var/home/core/zuul-output/0000755000175000017500000000000015073171434014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073201730015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004745060015073201722017702 0ustar rootrootOct 13 12:43:30 crc systemd[1]: Starting Kubernetes Kubelet... Oct 13 12:43:30 crc restorecon[4677]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 13 12:43:30 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 13 12:43:31 crc restorecon[4677]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 13 12:43:32 crc kubenswrapper[4678]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.350377 4678 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358724 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358757 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358767 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358777 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358786 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358796 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358805 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358813 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358822 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358833 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358842 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358850 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358858 4678 feature_gate.go:330] unrecognized feature gate: Example Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358866 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358882 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358890 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358898 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358906 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358916 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358924 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358933 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358941 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358949 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358959 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358969 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358978 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358986 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.358995 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359004 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359013 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359021 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359030 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359039 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359047 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359122 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359132 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359140 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359148 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359156 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359164 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359174 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359182 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359189 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359197 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359205 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359212 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359220 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359227 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359235 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359242 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359262 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359270 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359278 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359286 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359294 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359301 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359309 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359317 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359329 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359339 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359349 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359357 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359365 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359374 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359382 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359389 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359397 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359404 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359412 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359419 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.359429 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361551 4678 flags.go:64] FLAG: --address="0.0.0.0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361604 4678 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361624 4678 flags.go:64] FLAG: --anonymous-auth="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361637 4678 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361650 4678 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361661 4678 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361675 4678 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361686 4678 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361696 4678 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361706 4678 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361717 4678 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361727 4678 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361736 4678 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361745 4678 flags.go:64] FLAG: --cgroup-root="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361754 4678 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361763 4678 flags.go:64] FLAG: --client-ca-file="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361772 4678 flags.go:64] FLAG: --cloud-config="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361781 4678 flags.go:64] FLAG: --cloud-provider="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361790 4678 flags.go:64] FLAG: --cluster-dns="[]" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361801 4678 flags.go:64] FLAG: --cluster-domain="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361810 4678 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361820 4678 flags.go:64] FLAG: --config-dir="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361829 4678 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361839 4678 flags.go:64] FLAG: --container-log-max-files="5" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361851 4678 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361860 4678 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361871 4678 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361882 4678 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361893 4678 flags.go:64] FLAG: --contention-profiling="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361902 4678 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361911 4678 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361920 4678 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361929 4678 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361945 4678 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361954 4678 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361964 4678 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361975 4678 flags.go:64] FLAG: --enable-load-reader="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361985 4678 flags.go:64] FLAG: --enable-server="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.361994 4678 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362008 4678 flags.go:64] FLAG: --event-burst="100" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362017 4678 flags.go:64] FLAG: --event-qps="50" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362027 4678 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362036 4678 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362046 4678 flags.go:64] FLAG: --eviction-hard="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362086 4678 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362095 4678 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362104 4678 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362114 4678 flags.go:64] FLAG: --eviction-soft="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362123 4678 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362132 4678 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362142 4678 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362151 4678 flags.go:64] FLAG: --experimental-mounter-path="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362160 4678 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362169 4678 flags.go:64] FLAG: --fail-swap-on="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362178 4678 flags.go:64] FLAG: --feature-gates="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362189 4678 flags.go:64] FLAG: --file-check-frequency="20s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362199 4678 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362209 4678 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362218 4678 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362227 4678 flags.go:64] FLAG: --healthz-port="10248" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362237 4678 flags.go:64] FLAG: --help="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362246 4678 flags.go:64] FLAG: --hostname-override="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362255 4678 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362264 4678 flags.go:64] FLAG: --http-check-frequency="20s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362274 4678 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362283 4678 flags.go:64] FLAG: --image-credential-provider-config="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362292 4678 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362301 4678 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362310 4678 flags.go:64] FLAG: --image-service-endpoint="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362319 4678 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362328 4678 flags.go:64] FLAG: --kube-api-burst="100" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362337 4678 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362348 4678 flags.go:64] FLAG: --kube-api-qps="50" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362358 4678 flags.go:64] FLAG: --kube-reserved="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362367 4678 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362376 4678 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362385 4678 flags.go:64] FLAG: --kubelet-cgroups="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362393 4678 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362403 4678 flags.go:64] FLAG: --lock-file="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362411 4678 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362420 4678 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362429 4678 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362444 4678 flags.go:64] FLAG: --log-json-split-stream="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362453 4678 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362462 4678 flags.go:64] FLAG: --log-text-split-stream="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362471 4678 flags.go:64] FLAG: --logging-format="text" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362480 4678 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362489 4678 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362499 4678 flags.go:64] FLAG: --manifest-url="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362507 4678 flags.go:64] FLAG: --manifest-url-header="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362520 4678 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362530 4678 flags.go:64] FLAG: --max-open-files="1000000" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362541 4678 flags.go:64] FLAG: --max-pods="110" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362550 4678 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362559 4678 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362568 4678 flags.go:64] FLAG: --memory-manager-policy="None" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362577 4678 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362587 4678 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362599 4678 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362608 4678 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362637 4678 flags.go:64] FLAG: --node-status-max-images="50" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362647 4678 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362656 4678 flags.go:64] FLAG: --oom-score-adj="-999" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362665 4678 flags.go:64] FLAG: --pod-cidr="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362675 4678 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362689 4678 flags.go:64] FLAG: --pod-manifest-path="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362698 4678 flags.go:64] FLAG: --pod-max-pids="-1" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362707 4678 flags.go:64] FLAG: --pods-per-core="0" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362717 4678 flags.go:64] FLAG: --port="10250" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362727 4678 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362736 4678 flags.go:64] FLAG: --provider-id="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362745 4678 flags.go:64] FLAG: --qos-reserved="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362754 4678 flags.go:64] FLAG: --read-only-port="10255" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362763 4678 flags.go:64] FLAG: --register-node="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362772 4678 flags.go:64] FLAG: --register-schedulable="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362781 4678 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362797 4678 flags.go:64] FLAG: --registry-burst="10" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362806 4678 flags.go:64] FLAG: --registry-qps="5" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362815 4678 flags.go:64] FLAG: --reserved-cpus="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362826 4678 flags.go:64] FLAG: --reserved-memory="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362837 4678 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362846 4678 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362855 4678 flags.go:64] FLAG: --rotate-certificates="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362864 4678 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362874 4678 flags.go:64] FLAG: --runonce="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362883 4678 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362892 4678 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362901 4678 flags.go:64] FLAG: --seccomp-default="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362911 4678 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362920 4678 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362930 4678 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362939 4678 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362948 4678 flags.go:64] FLAG: --storage-driver-password="root" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362957 4678 flags.go:64] FLAG: --storage-driver-secure="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362966 4678 flags.go:64] FLAG: --storage-driver-table="stats" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362975 4678 flags.go:64] FLAG: --storage-driver-user="root" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362985 4678 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.362994 4678 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363003 4678 flags.go:64] FLAG: --system-cgroups="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363012 4678 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363026 4678 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363035 4678 flags.go:64] FLAG: --tls-cert-file="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363044 4678 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363076 4678 flags.go:64] FLAG: --tls-min-version="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363087 4678 flags.go:64] FLAG: --tls-private-key-file="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363096 4678 flags.go:64] FLAG: --topology-manager-policy="none" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363105 4678 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363114 4678 flags.go:64] FLAG: --topology-manager-scope="container" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363124 4678 flags.go:64] FLAG: --v="2" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363136 4678 flags.go:64] FLAG: --version="false" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363149 4678 flags.go:64] FLAG: --vmodule="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363160 4678 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.363170 4678 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363416 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363427 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363436 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363448 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363460 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363472 4678 feature_gate.go:330] unrecognized feature gate: Example Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363483 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363495 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363507 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363519 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363529 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363538 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363549 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363559 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363569 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363579 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363589 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363597 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363605 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363613 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363620 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363628 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363637 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363644 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363652 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363660 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363667 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363677 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363691 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363701 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363709 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363718 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363728 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363738 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363747 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363755 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363764 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363772 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363780 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363790 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363798 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363806 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363814 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363821 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363829 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363837 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363844 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363852 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363860 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363868 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363876 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363883 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363891 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363898 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363906 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363917 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363926 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363934 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363942 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363949 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363964 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363974 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363983 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363991 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.363999 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364009 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364016 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364024 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364032 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364039 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.364047 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.364107 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.374364 4678 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.374681 4678 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374791 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374806 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374816 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374824 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374834 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374846 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374858 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374868 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374877 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374885 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374895 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374904 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374913 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374921 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374929 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374938 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374947 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374955 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374963 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374972 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374981 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374989 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.374996 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375007 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375015 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375023 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375031 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375041 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375099 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375113 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375124 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375135 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375167 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375182 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375192 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375204 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375214 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375225 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375236 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375246 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375253 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375261 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375269 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375277 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375285 4678 feature_gate.go:330] unrecognized feature gate: Example Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375293 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375301 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375308 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375316 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375324 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375331 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375342 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375351 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375359 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375368 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375376 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375384 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375392 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375403 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375413 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375422 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375429 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375438 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375448 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375457 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375464 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375472 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375481 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375489 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375496 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375505 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.375517 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375733 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375744 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375752 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375761 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375769 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375778 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375786 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375796 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375805 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375813 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375821 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375829 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375837 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375845 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375853 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375861 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375870 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375878 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375890 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375900 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375908 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375917 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375925 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375933 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375942 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375950 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375958 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375966 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375976 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375986 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.375995 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376005 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376014 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376022 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376030 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376039 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376047 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376082 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376091 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376099 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376107 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376116 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376124 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376132 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376140 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376148 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376156 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376164 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376171 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376178 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376186 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376194 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376203 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376211 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376218 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376226 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376233 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376242 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376249 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376328 4678 feature_gate.go:330] unrecognized feature gate: Example Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376450 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376460 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376469 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376477 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376485 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376494 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376502 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376511 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376519 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376528 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.376539 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.376553 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.379255 4678 server.go:940] "Client rotation is on, will bootstrap in background" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.383878 4678 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.383983 4678 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.385704 4678 server.go:997] "Starting client certificate rotation" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.385739 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.385956 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-10 17:41:17.439097002 +0000 UTC Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.386108 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2140h57m45.052994573s for next certificate rotation Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.410718 4678 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.414895 4678 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.437160 4678 log.go:25] "Validated CRI v1 runtime API" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.474846 4678 log.go:25] "Validated CRI v1 image API" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.477363 4678 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.482077 4678 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-13-12-38-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.482102 4678 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.496876 4678 manager.go:217] Machine: {Timestamp:2025-10-13 12:43:32.493116904 +0000 UTC m=+0.577654808 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:bc86650e-4c93-41c1-ba1f-a59fdc4063ca BootID:4d85342c-260c-43e8-af13-3624a88b6f98 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7d:ad:7e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7d:ad:7e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ea:df:bb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:43:ec:0a Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a7:a7:9a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a0:2d:93 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:72:68:68:66:69:f4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:da:64:d0:b2:51:11 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497134 4678 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497262 4678 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497504 4678 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497656 4678 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497692 4678 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497874 4678 topology_manager.go:138] "Creating topology manager with none policy" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.497883 4678 container_manager_linux.go:303] "Creating device plugin manager" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.498330 4678 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.498363 4678 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.498547 4678 state_mem.go:36] "Initialized new in-memory state store" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.498627 4678 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.502940 4678 kubelet.go:418] "Attempting to sync node with API server" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.502962 4678 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.502984 4678 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.502997 4678 kubelet.go:324] "Adding apiserver pod source" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.503008 4678 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.506593 4678 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.507577 4678 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.509236 4678 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.509726 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.509744 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.509871 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.509793 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510697 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510719 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510726 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510733 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510744 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510752 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510760 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510771 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510779 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510787 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510796 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.510803 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.511601 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.511999 4678 server.go:1280] "Started kubelet" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.513225 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.513276 4678 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.513275 4678 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 13 12:43:32 crc systemd[1]: Started Kubernetes Kubelet. Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.513911 4678 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.515290 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.515350 4678 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.515553 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 00:58:09.386636298 +0000 UTC Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.515637 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1668h14m36.871006818s for next certificate rotation Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.515777 4678 server.go:460] "Adding debug handlers to kubelet server" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.516179 4678 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.516201 4678 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.516211 4678 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.516248 4678 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.516838 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.517195 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.518247 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="200ms" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.521201 4678 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e0d8f13caa84c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-13 12:43:32.5119755 +0000 UTC m=+0.596513384,LastTimestamp:2025-10-13 12:43:32.5119755 +0000 UTC m=+0.596513384,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.523878 4678 factory.go:55] Registering systemd factory Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.523917 4678 factory.go:221] Registration of the systemd container factory successfully Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.524572 4678 factory.go:153] Registering CRI-O factory Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.524643 4678 factory.go:221] Registration of the crio container factory successfully Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.524754 4678 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.524798 4678 factory.go:103] Registering Raw factory Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.524826 4678 manager.go:1196] Started watching for new ooms in manager Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.529816 4678 manager.go:319] Starting recovery of all containers Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534444 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534512 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534536 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534557 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534575 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534593 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534613 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534644 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534664 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534683 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534701 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534720 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534739 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534759 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534779 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534797 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534815 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534838 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534857 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534874 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534892 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.534912 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535014 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535033 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535080 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535109 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535133 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535152 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535169 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535190 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535210 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535261 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535281 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535300 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535317 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535336 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535393 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535415 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535441 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535460 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535478 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535495 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535514 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535532 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535551 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535569 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535587 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535606 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535624 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535642 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535661 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535678 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535705 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535726 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535746 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535765 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535784 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535806 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535826 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535845 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535865 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535883 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535901 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535919 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535937 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.535955 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536007 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536026 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536042 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536084 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536103 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536121 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536139 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536160 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536178 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536197 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536216 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536235 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536254 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536272 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536290 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536310 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536328 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536345 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536362 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536382 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536399 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536417 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536436 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536496 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536552 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536570 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536587 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536605 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536623 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536641 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536661 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536679 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536699 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536717 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536735 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536754 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536772 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536791 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536818 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536839 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536861 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536881 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536900 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536920 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536940 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536961 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.536983 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537003 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537024 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537042 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537087 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537108 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537126 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537144 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537163 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537184 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537203 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537221 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537241 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537260 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537279 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537298 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537317 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537335 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537354 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537372 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.537394 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539835 4678 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539883 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539908 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539925 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539942 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539954 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539970 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.539986 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540003 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540017 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540030 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540044 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540076 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540089 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540103 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540117 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540131 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540144 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540155 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540167 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540179 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540189 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540199 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540209 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540222 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540233 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540243 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540253 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540264 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540274 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540284 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540294 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540307 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540317 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540328 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540337 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540349 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540359 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540369 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540381 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540402 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540412 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540421 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540430 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540456 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540465 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540476 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540488 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540500 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540512 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540527 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540542 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540554 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540565 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540577 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540598 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540607 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540622 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540631 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540641 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540653 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540664 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540677 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540688 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540698 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540708 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540718 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540729 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540741 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540750 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540759 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540768 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540779 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540789 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540801 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540810 4678 reconstruct.go:97] "Volume reconstruction finished" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.540816 4678 reconciler.go:26] "Reconciler: start to sync state" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.557081 4678 manager.go:324] Recovery completed Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.573165 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.575079 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.575144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.575163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.576345 4678 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.576427 4678 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.576459 4678 state_mem.go:36] "Initialized new in-memory state store" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.587966 4678 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.590931 4678 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.591002 4678 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.591043 4678 kubelet.go:2335] "Starting kubelet main sync loop" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.591198 4678 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 13 12:43:32 crc kubenswrapper[4678]: W1013 12:43:32.598862 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.598961 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.602209 4678 policy_none.go:49] "None policy: Start" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.603358 4678 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.603405 4678 state_mem.go:35] "Initializing new in-memory state store" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.617268 4678 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.664287 4678 manager.go:334] "Starting Device Plugin manager" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.664672 4678 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.664714 4678 server.go:79] "Starting device plugin registration server" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.665407 4678 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.665440 4678 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.665709 4678 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.665803 4678 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.665810 4678 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.681628 4678 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.691428 4678 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.691763 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.694175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.694239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.694256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.694458 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.695300 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.695363 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696643 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696853 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696892 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.696954 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698522 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698724 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.698789 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700481 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700777 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.700831 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701465 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701483 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701740 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701794 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.701815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.702870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.702907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.702923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.719683 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="400ms" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744536 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744627 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744668 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744704 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744734 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744761 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744804 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744832 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744883 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.744946 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.745017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.745049 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.745102 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.745132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.745167 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.765876 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.767152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.767202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.767223 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.767255 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.767848 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.846741 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.846819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847217 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847162 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847290 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847317 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847334 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847353 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847371 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847388 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847409 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847427 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847459 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847459 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847477 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847467 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847515 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847544 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847455 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847561 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847424 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847517 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847488 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847384 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847534 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847616 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847650 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.847948 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.968922 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.970392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.970448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.970465 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:32 crc kubenswrapper[4678]: I1013 12:43:32.970497 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:32 crc kubenswrapper[4678]: E1013 12:43:32.971141 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.031029 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.039230 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.057199 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.062881 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.092192 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-1a3be9fbbcdd63beb44e5b7397a1f05c44c7393e439d52d3c6743b65b3337cf3 WatchSource:0}: Error finding container 1a3be9fbbcdd63beb44e5b7397a1f05c44c7393e439d52d3c6743b65b3337cf3: Status 404 returned error can't find the container with id 1a3be9fbbcdd63beb44e5b7397a1f05c44c7393e439d52d3c6743b65b3337cf3 Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.093516 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.094350 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-b491c2a51619fcaa10bc5730d7b430f8100de8bba10c2e46381bdc99c2d2066d WatchSource:0}: Error finding container b491c2a51619fcaa10bc5730d7b430f8100de8bba10c2e46381bdc99c2d2066d: Status 404 returned error can't find the container with id b491c2a51619fcaa10bc5730d7b430f8100de8bba10c2e46381bdc99c2d2066d Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.105206 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-564cbc74424a8c46025ac1b6a5fe87991040364bb21f1addb02e116fb783ff78 WatchSource:0}: Error finding container 564cbc74424a8c46025ac1b6a5fe87991040364bb21f1addb02e116fb783ff78: Status 404 returned error can't find the container with id 564cbc74424a8c46025ac1b6a5fe87991040364bb21f1addb02e116fb783ff78 Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.115992 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-153d63d0f63be4fabc4c3f2f4b36c940b81afeb51cf3be7509e1066117e2fb68 WatchSource:0}: Error finding container 153d63d0f63be4fabc4c3f2f4b36c940b81afeb51cf3be7509e1066117e2fb68: Status 404 returned error can't find the container with id 153d63d0f63be4fabc4c3f2f4b36c940b81afeb51cf3be7509e1066117e2fb68 Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.121018 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="800ms" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.371673 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.372795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.372825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.372836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.372859 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.373311 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.381015 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.381140 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.514307 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.531152 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.531247 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.595568 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0f7dd68d7d8c59eb99d53d03072ae1abbd62b0bc9a15fe308c8253da295a21f9"} Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.596749 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"564cbc74424a8c46025ac1b6a5fe87991040364bb21f1addb02e116fb783ff78"} Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.597903 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1a3be9fbbcdd63beb44e5b7397a1f05c44c7393e439d52d3c6743b65b3337cf3"} Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.599021 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b491c2a51619fcaa10bc5730d7b430f8100de8bba10c2e46381bdc99c2d2066d"} Oct 13 12:43:33 crc kubenswrapper[4678]: I1013 12:43:33.599678 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"153d63d0f63be4fabc4c3f2f4b36c940b81afeb51cf3be7509e1066117e2fb68"} Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.696108 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.696247 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:33 crc kubenswrapper[4678]: W1013 12:43:33.832538 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.832659 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:33 crc kubenswrapper[4678]: E1013 12:43:33.921813 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="1.6s" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.173528 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.175763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.176235 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.176254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.176293 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:34 crc kubenswrapper[4678]: E1013 12:43:34.176897 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.514028 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.604367 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d" exitCode=0 Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.604444 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.604515 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.605727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.605796 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.605822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.607289 4678 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="73781957fc5d35902182dbd31ec90b4254d867825c76296ddae10d6f96ddb333" exitCode=0 Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.607344 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.607381 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"73781957fc5d35902182dbd31ec90b4254d867825c76296ddae10d6f96ddb333"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.607998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.608018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.608026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.609702 4678 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100" exitCode=0 Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.609755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.609919 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.611254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.611276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.611284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.614697 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.614749 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.614762 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.614769 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.614880 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.620576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.620628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.620640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.623103 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557" exitCode=0 Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.623134 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557"} Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.623331 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.624895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.624962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.624984 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.628863 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.630281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.630343 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:34 crc kubenswrapper[4678]: I1013 12:43:34.630369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.514786 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:35 crc kubenswrapper[4678]: E1013 12:43:35.522561 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="3.2s" Oct 13 12:43:35 crc kubenswrapper[4678]: W1013 12:43:35.522570 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Oct 13 12:43:35 crc kubenswrapper[4678]: E1013 12:43:35.522636 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.630107 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.630223 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.630249 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.630268 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.634076 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940" exitCode=0 Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.634170 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.634318 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.635861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.635898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.635909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.640543 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"de74300447ee49ec05aba18d895058af13c1ca2fc4ac500f5249d7ccd163c8a6"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.640829 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.642452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.642813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643152 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643184 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643223 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0"} Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643227 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643235 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.643125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.644701 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.644725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.644734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.644755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.645829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.645871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.777312 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.783248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.783300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.783311 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:35 crc kubenswrapper[4678]: I1013 12:43:35.783335 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:35 crc kubenswrapper[4678]: E1013 12:43:35.784204 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.654782 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4"} Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.654930 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.656607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.656655 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.656676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.659907 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735" exitCode=0 Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.660080 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735"} Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.660216 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.660291 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.660228 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.660855 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662173 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.662900 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:36 crc kubenswrapper[4678]: I1013 12:43:36.663030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.378893 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.379503 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.383136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.383200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.383219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.389933 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668550 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668609 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668661 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668594 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f"} Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668627 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da"} Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.668868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee"} Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.670902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.671958 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.672006 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.672022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:37 crc kubenswrapper[4678]: I1013 12:43:37.825551 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.531787 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631"} Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674802 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674811 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2"} Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674784 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674844 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674908 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.674977 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676334 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.676351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.984877 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.986731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.986779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.986800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:38 crc kubenswrapper[4678]: I1013 12:43:38.986833 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.642713 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.677199 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.677308 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678751 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:39 crc kubenswrapper[4678]: I1013 12:43:39.678768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:40 crc kubenswrapper[4678]: I1013 12:43:40.293850 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:40 crc kubenswrapper[4678]: I1013 12:43:40.294101 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:40 crc kubenswrapper[4678]: I1013 12:43:40.295472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:40 crc kubenswrapper[4678]: I1013 12:43:40.295525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:40 crc kubenswrapper[4678]: I1013 12:43:40.295543 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.532872 4678 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.532991 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.535372 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.535654 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.537333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.537382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:41 crc kubenswrapper[4678]: I1013 12:43:41.537400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:42 crc kubenswrapper[4678]: E1013 12:43:42.681790 4678 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 13 12:43:43 crc kubenswrapper[4678]: I1013 12:43:43.211862 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 13 12:43:43 crc kubenswrapper[4678]: I1013 12:43:43.212457 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:43 crc kubenswrapper[4678]: I1013 12:43:43.214161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:43 crc kubenswrapper[4678]: I1013 12:43:43.214229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:43 crc kubenswrapper[4678]: I1013 12:43:43.214248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.157040 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.157266 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.158705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.158760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.158782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.164099 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.694205 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.695467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.695641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:44 crc kubenswrapper[4678]: I1013 12:43:44.695837 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.171328 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.171674 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.173230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.173282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.173302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:46 crc kubenswrapper[4678]: W1013 12:43:46.200791 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.200955 4678 trace.go:236] Trace[1477580219]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 12:43:36.199) (total time: 10001ms): Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[1477580219]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:43:46.200) Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[1477580219]: [10.001584827s] [10.001584827s] END Oct 13 12:43:46 crc kubenswrapper[4678]: E1013 12:43:46.201003 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 13 12:43:46 crc kubenswrapper[4678]: W1013 12:43:46.387248 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.387369 4678 trace.go:236] Trace[199388493]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 12:43:36.385) (total time: 10001ms): Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[199388493]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:43:46.387) Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[199388493]: [10.00146286s] [10.00146286s] END Oct 13 12:43:46 crc kubenswrapper[4678]: E1013 12:43:46.387403 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.515531 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 13 12:43:46 crc kubenswrapper[4678]: W1013 12:43:46.806007 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.806173 4678 trace.go:236] Trace[1579287477]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 12:43:36.804) (total time: 10001ms): Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[1579287477]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (12:43:46.805) Oct 13 12:43:46 crc kubenswrapper[4678]: Trace[1579287477]: [10.001757093s] [10.001757093s] END Oct 13 12:43:46 crc kubenswrapper[4678]: E1013 12:43:46.806206 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.846725 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.846790 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.862915 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 13 12:43:46 crc kubenswrapper[4678]: I1013 12:43:46.863004 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 13 12:43:50 crc kubenswrapper[4678]: I1013 12:43:50.747332 4678 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 13 12:43:50 crc kubenswrapper[4678]: I1013 12:43:50.952540 4678 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.068179 4678 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.516036 4678 apiserver.go:52] "Watching apiserver" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.521035 4678 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.521567 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.522529 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.522783 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.522965 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.522971 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.523117 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.523164 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.523458 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.523530 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.523624 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.526671 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527294 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527606 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527745 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527331 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527441 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.527485 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.529974 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.529976 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.532714 4678 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.532992 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.541313 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.549906 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.562351 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.568729 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.590206 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.602707 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.613892 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.616871 4678 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.625172 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.634744 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.647464 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.658346 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.669611 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.712200 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.719313 4678 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.845998 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.847577 4678 trace.go:236] Trace[1478783805]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Oct-2025 12:43:41.152) (total time: 10694ms): Oct 13 12:43:51 crc kubenswrapper[4678]: Trace[1478783805]: ---"Objects listed" error: 10694ms (12:43:51.847) Oct 13 12:43:51 crc kubenswrapper[4678]: Trace[1478783805]: [10.69481075s] [10.69481075s] END Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.847605 4678 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.850009 4678 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.850002 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.926676 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.944842 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950524 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950582 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950615 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950642 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950668 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950691 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950714 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950736 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950761 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950788 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950817 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950843 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950866 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950891 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950913 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950936 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950961 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.950985 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951036 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951082 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951105 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951127 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951150 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951214 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951241 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951267 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951312 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951341 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951365 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951362 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951390 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951435 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951461 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951484 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951506 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951533 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951562 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951588 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951618 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951644 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951694 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951723 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951725 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951746 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951776 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951799 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951857 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951878 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951902 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951924 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951947 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951971 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.951995 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952018 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952042 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952085 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952113 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952138 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952163 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952189 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952217 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952240 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952267 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952294 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952343 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952340 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952370 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952399 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952426 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952453 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952480 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952507 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952508 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952507 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952532 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952694 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952768 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952780 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952802 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952834 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952889 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.952955 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953008 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953110 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953121 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953173 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953228 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953369 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953380 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953436 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953449 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953511 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953724 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953786 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953840 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953893 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.953953 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954002 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954050 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954135 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954184 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954235 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954290 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954344 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954394 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954456 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954505 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954555 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954606 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954653 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954702 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954749 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954795 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954909 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.954962 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955017 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955114 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955166 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955219 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955270 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955328 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955379 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955431 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955465 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955482 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955535 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955594 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955654 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955714 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955766 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955815 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955857 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955902 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955907 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.955961 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956000 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956037 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956084 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956113 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956151 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956172 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956186 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956222 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956333 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956389 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956429 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956470 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956503 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956540 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956578 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956613 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956648 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956168 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956740 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956780 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.956895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957178 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.957218 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:43:52.457190052 +0000 UTC m=+20.541728176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957265 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957334 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957617 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957682 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957730 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957827 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957847 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957873 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957890 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957916 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.957959 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958000 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958024 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958097 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958139 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958177 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958216 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958253 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958336 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958366 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958835 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.958375 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.959983 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960037 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960111 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960149 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960186 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960221 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960258 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960294 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960333 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960331 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960385 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960505 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960588 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960759 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960809 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960851 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960903 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960948 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.960999 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961183 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961244 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961271 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961407 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961445 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961486 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961523 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961561 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961601 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961638 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961747 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961792 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961824 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961859 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961897 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961922 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.961947 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962018 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962086 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962127 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962201 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962228 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962254 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962286 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962323 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962353 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962379 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962409 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962446 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962500 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962640 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962666 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962687 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962707 4678 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962728 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962750 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962771 4678 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962790 4678 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962786 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962813 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962836 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962846 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962856 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962898 4678 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962917 4678 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962936 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962952 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962967 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.962983 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963026 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963041 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963075 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963092 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963164 4678 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963181 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963196 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963324 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.963430 4678 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.964027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.964355 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.964485 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.964532 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.964845 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965045 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965121 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965184 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965223 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965559 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965777 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965953 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.965995 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966163 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966432 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966462 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966720 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966722 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966883 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.966978 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967250 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967271 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967309 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967032 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967330 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967368 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967429 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967469 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967518 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967575 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967651 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.967837 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.968503 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.968928 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.969263 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.969424 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.969294 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970025 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970273 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970548 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970628 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970741 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970793 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.970726 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.971300 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.972139 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.972549 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.972809 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.972769 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.972862 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.973257 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:52.472818746 +0000 UTC m=+20.557356830 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.972579 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.973530 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.973255 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.973783 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.973985 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.974609 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.975946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.976492 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.976608 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:51 crc kubenswrapper[4678]: E1013 12:43:51.976705 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:52.476679783 +0000 UTC m=+20.561217687 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.977300 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.977744 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.977803 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.978088 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.978938 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.979042 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.979918 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.980327 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.980400 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.980653 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.980897 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.981169 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.981368 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.987272 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.987504 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.987841 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.988406 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.989203 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.989203 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.989284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.989777 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.989980 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.990049 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.990349 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.991119 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.991175 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.991664 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.991906 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.992194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.992401 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.992803 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.992937 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.993429 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.993517 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.993565 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.993783 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.994187 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.994248 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.994264 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.994285 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.994753 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995081 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995554 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995619 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995661 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995674 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.995816 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.997255 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.997518 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.997403 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.997398 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998052 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998052 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998145 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998264 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998399 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998411 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998601 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998708 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.998756 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999110 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999155 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999180 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999192 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999225 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999273 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:51 crc kubenswrapper[4678]: I1013 12:43:51.999557 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:51.998867 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000379 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000425 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000441 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000469 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000534 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000829 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000973 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.000982 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001437 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001002 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001652 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001723 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001190 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001281 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001369 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001445 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.001023 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.002247 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.002829 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.003186 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.003223 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.003621 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.003645 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.003659 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.003925 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.003978 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.004136 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:52.504118024 +0000 UTC m=+20.588655908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.004306 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.004397 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.004401 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.004517 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.005091 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.005790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.005852 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.007118 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.007174 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.007635 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.007683 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.007857 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.008042 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.008171 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.008991 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.009283 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018007 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.018166 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.018179 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.018190 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.018237 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:52.51821972 +0000 UTC m=+20.602757604 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018182 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018478 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018613 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018676 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.018719 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.020282 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.029027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.031323 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.041672 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.053805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.055987 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063617 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063700 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063775 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063790 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063808 4678 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063821 4678 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063832 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063844 4678 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063856 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063868 4678 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063879 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063890 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063901 4678 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063912 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063923 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063935 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063946 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063956 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063969 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.063982 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064023 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064036 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064051 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064081 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064093 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064104 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064115 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064126 4678 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064136 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064148 4678 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064159 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064171 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064183 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064194 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064205 4678 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064215 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064226 4678 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064237 4678 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064247 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064258 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064269 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064281 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064293 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064305 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064316 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064345 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064359 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064372 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064384 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064395 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064407 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064424 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064435 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064446 4678 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064457 4678 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064468 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064478 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064488 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064499 4678 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064510 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064521 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064532 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064543 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064556 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064571 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064582 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064592 4678 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064603 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064614 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064625 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064635 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064645 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064657 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064667 4678 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064678 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064689 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064703 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064716 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064729 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064741 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064752 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064763 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064773 4678 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064785 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064798 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064814 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064828 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064840 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064860 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064874 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064886 4678 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064899 4678 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064911 4678 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064921 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064932 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064942 4678 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064952 4678 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064963 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064974 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064984 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.064995 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065006 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065017 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065028 4678 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065039 4678 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065078 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065091 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065103 4678 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065114 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065126 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065137 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065148 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065160 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065172 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065183 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065193 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065206 4678 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065216 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065227 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065239 4678 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065250 4678 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065263 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065275 4678 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065287 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065298 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065310 4678 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065322 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065335 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065346 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065357 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065369 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065380 4678 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065392 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065403 4678 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065414 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065426 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065437 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065452 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065463 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065474 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065485 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065498 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065509 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065520 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065531 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065541 4678 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065552 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065564 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065575 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065589 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065601 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065613 4678 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065624 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065634 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065646 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065657 4678 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065668 4678 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065681 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065692 4678 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065703 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065714 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065725 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.065735 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067297 4678 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067433 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067454 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067467 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067557 4678 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.067565 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.074864 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.075296 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.075363 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.075594 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.076315 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.080447 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.147188 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.167238 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.168400 4678 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.168423 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.168432 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.177146 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 13 12:43:52 crc kubenswrapper[4678]: W1013 12:43:52.190691 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-674fa470190a41d575993bca25a1a1f2248ce53965d9108f2adfcc40a0237e3d WatchSource:0}: Error finding container 674fa470190a41d575993bca25a1a1f2248ce53965d9108f2adfcc40a0237e3d: Status 404 returned error can't find the container with id 674fa470190a41d575993bca25a1a1f2248ce53965d9108f2adfcc40a0237e3d Oct 13 12:43:52 crc kubenswrapper[4678]: W1013 12:43:52.193595 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-c22d8f8ab623dc494f055eded866c94ef871dc9cad2925f310bb14e0a01591ad WatchSource:0}: Error finding container c22d8f8ab623dc494f055eded866c94ef871dc9cad2925f310bb14e0a01591ad: Status 404 returned error can't find the container with id c22d8f8ab623dc494f055eded866c94ef871dc9cad2925f310bb14e0a01591ad Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.470815 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.470998 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:43:53.470957021 +0000 UTC m=+21.555494915 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.572662 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.572713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.572737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.572762 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572896 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572939 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572939 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572980 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572908 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572994 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.573005 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.573019 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.572988 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:53.572966921 +0000 UTC m=+21.657504905 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.573088 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:53.573071574 +0000 UTC m=+21.657609468 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.573104 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:53.573096165 +0000 UTC m=+21.657634059 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: E1013 12:43:52.573117 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:53.573111035 +0000 UTC m=+21.657648929 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.596389 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.597035 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.597736 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.598413 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.598971 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.600245 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.600764 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.601725 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.602451 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.603369 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.603849 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.604806 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.605262 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.605758 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.606610 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.607104 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.608091 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.608448 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.608966 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.609880 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.610307 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.611269 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.611753 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.611860 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.612779 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.613188 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.613734 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.614918 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.615373 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.616396 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.616812 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.617674 4678 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.617792 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.619866 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.620673 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.621050 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.622500 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.623201 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.624017 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.624736 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.625759 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.626327 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.627211 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.627377 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.627972 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.629567 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.630040 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.630900 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.631434 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.632503 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.633014 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.633938 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.634441 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.635345 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.635912 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.635902 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.636372 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.650988 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.660832 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.671607 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.684298 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.715134 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c22d8f8ab623dc494f055eded866c94ef871dc9cad2925f310bb14e0a01591ad"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.716699 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.716826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.716907 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"674fa470190a41d575993bca25a1a1f2248ce53965d9108f2adfcc40a0237e3d"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.719153 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.719352 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fa7319e328fe30f5d869601abbcc6b7dfeb0d32b2e14526cadcfe9b4ad534c17"} Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.729195 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.737309 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.748312 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.761091 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.777324 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.814905 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.854428 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.876011 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.887103 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.902352 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.913477 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.925706 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.936459 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:52 crc kubenswrapper[4678]: I1013 12:43:52.950403 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.478928 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.479068 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:43:55.479013697 +0000 UTC m=+23.563551581 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.579939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.580192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580119 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580245 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580260 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580317 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:55.58029871 +0000 UTC m=+23.664836594 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580529 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580671 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:55.580643648 +0000 UTC m=+23.665181542 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.580702 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.580763 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580845 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580944 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580976 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.581010 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:55.580999297 +0000 UTC m=+23.665537191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.580881 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.581090 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:55.581076929 +0000 UTC m=+23.665615053 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.591623 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.591623 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.591780 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.591732 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.591838 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:43:53 crc kubenswrapper[4678]: E1013 12:43:53.591996 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.746310 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-925jf"] Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.746652 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.749515 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-nlfg7"] Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.749948 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.750237 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-bxtwv"] Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.750246 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.750685 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.751687 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.752175 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.753466 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.753604 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.753752 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.753862 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.753979 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.754732 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.754857 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.755089 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.755193 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.757073 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9mbz2"] Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.765420 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.771323 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.773784 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.773783 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.775495 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.794645 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.810985 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.823128 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.835295 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.853665 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.864949 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.880423 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.882931 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-hostroot\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.882969 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-rootfs\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.882991 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-os-release\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883014 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-socket-dir-parent\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-bin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883182 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-system-cni-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883310 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883411 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxp2g\" (UniqueName: \"kubernetes.io/projected/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-kube-api-access-xxp2g\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883470 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6sdx\" (UniqueName: \"kubernetes.io/projected/6539fec7-6f19-4758-8da3-13a6f978d5a6-kube-api-access-j6sdx\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883521 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-cnibin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883558 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-netns\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883590 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-kubelet\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzbsl\" (UniqueName: \"kubernetes.io/projected/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-kube-api-access-nzbsl\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883703 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-cni-binary-copy\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-daemon-config\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w95kl\" (UniqueName: \"kubernetes.io/projected/21d5191d-621a-4416-b635-e3459b3cfe8d-kube-api-access-w95kl\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883818 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-proxy-tls\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883856 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-conf-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883888 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883923 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.883958 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-os-release\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884006 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6539fec7-6f19-4758-8da3-13a6f978d5a6-hosts-file\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884037 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cnibin\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884102 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-etc-kubernetes\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884147 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-k8s-cni-cncf-io\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884198 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-multus-certs\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884237 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884273 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-mcd-auth-proxy-config\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884326 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-system-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.884358 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-multus\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.892330 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.907514 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.923475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.941487 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.954750 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.973104 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984855 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-os-release\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984900 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-hostroot\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984918 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-rootfs\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984934 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-socket-dir-parent\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984967 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-bin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.984984 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-system-cni-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985001 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985017 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6sdx\" (UniqueName: \"kubernetes.io/projected/6539fec7-6f19-4758-8da3-13a6f978d5a6-kube-api-access-j6sdx\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985032 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxp2g\" (UniqueName: \"kubernetes.io/projected/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-kube-api-access-xxp2g\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985038 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-hostroot\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985070 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-cnibin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-rootfs\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-socket-dir-parent\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985202 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-netns\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985224 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-system-cni-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985235 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-kubelet\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985263 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzbsl\" (UniqueName: \"kubernetes.io/projected/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-kube-api-access-nzbsl\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985317 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-cni-binary-copy\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-daemon-config\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985362 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w95kl\" (UniqueName: \"kubernetes.io/projected/21d5191d-621a-4416-b635-e3459b3cfe8d-kube-api-access-w95kl\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985386 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-proxy-tls\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985412 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-conf-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985434 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985479 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6539fec7-6f19-4758-8da3-13a6f978d5a6-hosts-file\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985508 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cnibin\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985529 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-os-release\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985556 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-etc-kubernetes\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985582 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-k8s-cni-cncf-io\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985631 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-multus-certs\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985662 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985687 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-mcd-auth-proxy-config\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985748 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-system-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985770 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-multus\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985820 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985850 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-multus\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-os-release\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986240 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-k8s-cni-cncf-io\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/6539fec7-6f19-4758-8da3-13a6f978d5a6-hosts-file\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986316 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cnibin\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986359 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-os-release\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986392 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-etc-kubernetes\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986394 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986437 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-multus-certs\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985119 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-cnibin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986579 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-run-netns\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986576 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-kubelet\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.985208 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-host-var-lib-cni-bin\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-system-cni-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.986967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-conf-dir\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.987266 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.987281 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-cni-binary-copy\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.987346 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.987384 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-mcd-auth-proxy-config\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.987628 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/21d5191d-621a-4416-b635-e3459b3cfe8d-multus-daemon-config\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.994658 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-proxy-tls\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:53 crc kubenswrapper[4678]: I1013 12:43:53.996916 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:53Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.007502 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6sdx\" (UniqueName: \"kubernetes.io/projected/6539fec7-6f19-4758-8da3-13a6f978d5a6-kube-api-access-j6sdx\") pod \"node-resolver-925jf\" (UID: \"6539fec7-6f19-4758-8da3-13a6f978d5a6\") " pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.011740 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w95kl\" (UniqueName: \"kubernetes.io/projected/21d5191d-621a-4416-b635-e3459b3cfe8d-kube-api-access-w95kl\") pod \"multus-nlfg7\" (UID: \"21d5191d-621a-4416-b635-e3459b3cfe8d\") " pod="openshift-multus/multus-nlfg7" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.013155 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxp2g\" (UniqueName: \"kubernetes.io/projected/8c6e2144-3790-4a4a-b47c-5b3459dacdf9-kube-api-access-xxp2g\") pod \"machine-config-daemon-bxtwv\" (UID: \"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\") " pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.013798 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzbsl\" (UniqueName: \"kubernetes.io/projected/2dee984d-00a7-4d89-a39f-1cfaac6bb2f6-kube-api-access-nzbsl\") pod \"multus-additional-cni-plugins-9mbz2\" (UID: \"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\") " pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.015246 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.030870 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.041466 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.050941 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.066212 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-925jf" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.074567 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nlfg7" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.080613 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.089968 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" Oct 13 12:43:54 crc kubenswrapper[4678]: W1013 12:43:54.095540 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c6e2144_3790_4a4a_b47c_5b3459dacdf9.slice/crio-b2325a08aff2a0dcf95f23d10aa019c80dbe9254bf83b36be3a4579691f32e59 WatchSource:0}: Error finding container b2325a08aff2a0dcf95f23d10aa019c80dbe9254bf83b36be3a4579691f32e59: Status 404 returned error can't find the container with id b2325a08aff2a0dcf95f23d10aa019c80dbe9254bf83b36be3a4579691f32e59 Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.149240 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hsx89"] Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.149978 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.151975 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.152181 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.152322 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.158449 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.158897 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.159142 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.159340 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.187601 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.206637 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.223360 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.239185 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.252278 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.265445 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.285906 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288676 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288712 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288731 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288772 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288794 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4tf8\" (UniqueName: \"kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.288813 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.289936 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290028 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290069 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290091 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290125 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290168 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290189 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290248 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290304 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290323 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290346 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.290373 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.302031 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.318246 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.333227 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.345834 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.357550 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391431 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391481 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391498 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391516 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391536 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391571 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391572 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391634 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391637 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391590 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391670 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391681 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391753 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4tf8\" (UniqueName: \"kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391772 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391810 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391825 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391845 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391862 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391882 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391898 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391918 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391936 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.391992 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392015 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392035 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392332 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392391 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392434 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392762 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392791 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392795 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392823 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.392871 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.397799 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.418032 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4tf8\" (UniqueName: \"kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8\") pod \"ovnkube-node-hsx89\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.483936 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:43:54 crc kubenswrapper[4678]: W1013 12:43:54.496777 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-40bad1607225e7f28937bdc9d263bc6e82cd37a6d4e5bb243b3834c213745b20 WatchSource:0}: Error finding container 40bad1607225e7f28937bdc9d263bc6e82cd37a6d4e5bb243b3834c213745b20: Status 404 returned error can't find the container with id 40bad1607225e7f28937bdc9d263bc6e82cd37a6d4e5bb243b3834c213745b20 Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.725445 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerStarted","Data":"43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.725489 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerStarted","Data":"8bfd8a16c4d26d97abd1dcdf1cd2ed4c3a606ed0add722b499513a8ef8690c4d"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.727672 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd" exitCode=0 Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.727736 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.727755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerStarted","Data":"392c802a68bb8f140403f9ab04c62810ccc53afb832c6993a0ca7f5cd054a426"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.730020 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-925jf" event={"ID":"6539fec7-6f19-4758-8da3-13a6f978d5a6","Type":"ContainerStarted","Data":"3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.730078 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-925jf" event={"ID":"6539fec7-6f19-4758-8da3-13a6f978d5a6","Type":"ContainerStarted","Data":"bc52e99f225f8a9b0229fb77c64ccea40e4d86e700e16a0bbb65b3b71f07e219"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.731543 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.732828 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" exitCode=0 Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.732881 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.732900 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"40bad1607225e7f28937bdc9d263bc6e82cd37a6d4e5bb243b3834c213745b20"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.735291 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.735333 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.735347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"b2325a08aff2a0dcf95f23d10aa019c80dbe9254bf83b36be3a4579691f32e59"} Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.744040 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.766468 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.777234 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.790667 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.804921 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.815748 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.829502 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.847563 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.859945 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.877592 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.902669 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.934820 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.954025 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:54 crc kubenswrapper[4678]: I1013 12:43:54.988134 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:54Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.016253 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.029287 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.039763 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.053099 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.071918 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.084000 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.099462 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.114014 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.126929 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.139318 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.504503 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.504695 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:43:59.504667429 +0000 UTC m=+27.589205313 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.592431 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.592477 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.592445 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.592631 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.592726 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.592808 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.605325 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.605389 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.605447 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.605492 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605519 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605590 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605612 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605623 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605646 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605682 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605597 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:59.605573632 +0000 UTC m=+27.690111556 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605716 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605734 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:59.605719796 +0000 UTC m=+27.690257720 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605738 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605759 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:59.605747267 +0000 UTC m=+27.690285191 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:55 crc kubenswrapper[4678]: E1013 12:43:55.605809 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:43:59.605786468 +0000 UTC m=+27.690324382 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746298 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746362 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746383 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746401 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746418 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.746434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.748962 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b" exitCode=0 Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.749063 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b"} Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.765032 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.780634 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.795412 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.813317 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.830567 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.848510 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.863855 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.875633 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.896299 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.931861 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.945015 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:55 crc kubenswrapper[4678]: I1013 12:43:55.960530 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:55Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.059761 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-57x29"] Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.060201 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.063607 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.063828 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.063939 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.064371 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.075466 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.090853 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.109781 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.121488 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.135950 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.151160 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.165089 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.180490 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.196462 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.203671 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.212172 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.213017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f51fa35-1679-4abf-8d05-1070830d0f42-host\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.213073 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3f51fa35-1679-4abf-8d05-1070830d0f42-serviceca\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.213320 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7x5p\" (UniqueName: \"kubernetes.io/projected/3f51fa35-1679-4abf-8d05-1070830d0f42-kube-api-access-v7x5p\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.216943 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.218592 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.225776 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.238128 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.250662 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.263817 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.276992 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.286997 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.311042 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.314261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7x5p\" (UniqueName: \"kubernetes.io/projected/3f51fa35-1679-4abf-8d05-1070830d0f42-kube-api-access-v7x5p\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.314295 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f51fa35-1679-4abf-8d05-1070830d0f42-host\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.314324 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3f51fa35-1679-4abf-8d05-1070830d0f42-serviceca\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.314411 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3f51fa35-1679-4abf-8d05-1070830d0f42-host\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.315161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3f51fa35-1679-4abf-8d05-1070830d0f42-serviceca\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.321783 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.332991 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.338699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7x5p\" (UniqueName: \"kubernetes.io/projected/3f51fa35-1679-4abf-8d05-1070830d0f42-kube-api-access-v7x5p\") pod \"node-ca-57x29\" (UID: \"3f51fa35-1679-4abf-8d05-1070830d0f42\") " pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.345498 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.358633 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.368177 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.374451 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-57x29" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.381832 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: W1013 12:43:56.389072 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f51fa35_1679_4abf_8d05_1070830d0f42.slice/crio-671a1e92bee9fa2dd8dd1ca7bf0a33c6151821d34d403651329807cc2a6e627d WatchSource:0}: Error finding container 671a1e92bee9fa2dd8dd1ca7bf0a33c6151821d34d403651329807cc2a6e627d: Status 404 returned error can't find the container with id 671a1e92bee9fa2dd8dd1ca7bf0a33c6151821d34d403651329807cc2a6e627d Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.402505 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.415104 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.428898 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.441999 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.760466 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-57x29" event={"ID":"3f51fa35-1679-4abf-8d05-1070830d0f42","Type":"ContainerStarted","Data":"52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38"} Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.760544 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-57x29" event={"ID":"3f51fa35-1679-4abf-8d05-1070830d0f42","Type":"ContainerStarted","Data":"671a1e92bee9fa2dd8dd1ca7bf0a33c6151821d34d403651329807cc2a6e627d"} Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.764510 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37" exitCode=0 Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.764918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37"} Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.784353 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.798957 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.812847 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.835105 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.853887 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.867875 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.884675 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.907787 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.923832 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.942390 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.959175 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:56 crc kubenswrapper[4678]: I1013 12:43:56.975648 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:56Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.005818 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.030948 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.057016 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.111204 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.130083 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.140902 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.157393 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.170702 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.184669 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.197809 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.208670 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.220513 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.234428 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.245527 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.260891 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.310517 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.592123 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.592181 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:57 crc kubenswrapper[4678]: E1013 12:43:57.592241 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.592124 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:57 crc kubenswrapper[4678]: E1013 12:43:57.592369 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:43:57 crc kubenswrapper[4678]: E1013 12:43:57.592497 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.775019 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.786821 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2" exitCode=0 Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.786876 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2"} Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.818250 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.842475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.864807 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.880083 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.896726 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.911999 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.929729 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.941106 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.952323 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.969239 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:57 crc kubenswrapper[4678]: I1013 12:43:57.985436 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:57Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.006534 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.018722 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.029491 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.251003 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.254420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.254494 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.254519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.254707 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.266535 4678 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.266947 4678 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.268371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.268439 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.268451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.268508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.268527 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.290123 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.321559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.321620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.321637 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.321660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.321679 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.337458 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.342034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.342108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.342124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.342172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.342218 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.366487 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.370487 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.370525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.370536 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.370553 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.370565 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.386210 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.392441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.392508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.392518 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.392560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.392575 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.414510 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: E1013 12:43:58.414649 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.416691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.416724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.416736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.416753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.416766 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.524649 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.525421 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.525493 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.525581 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.525674 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.540184 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.555663 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.560615 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.560973 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.576204 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.593143 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.610681 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.626155 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.628104 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.628152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.628173 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.628200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.628219 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.641984 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.658143 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.679519 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.703410 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.731107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.731154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.731168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.731186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.731199 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.734701 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.747678 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.767121 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.778974 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.789872 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.791623 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec" exitCode=0 Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.792123 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.817775 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.832071 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.835352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.835384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.835396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.835418 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.835431 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.843849 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.856554 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.874682 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.885994 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.899947 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.917549 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.927862 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.937846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.937884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.937895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.937909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.937919 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:58Z","lastTransitionTime":"2025-10-13T12:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.940407 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.955511 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.968346 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:58 crc kubenswrapper[4678]: I1013 12:43:58.999242 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:58Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.036748 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.040794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.040836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.040847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.041046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.041074 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.085462 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.116280 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.143682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.143736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.143745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.143763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.143776 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.171030 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.203259 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.243349 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.246011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.246071 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.246086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.246106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.246120 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.277976 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.319367 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.349042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.349108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.349121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.349137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.349150 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.360879 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.406159 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.451919 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.451957 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.451965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.451979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.451989 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.457102 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.482485 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.522277 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.546199 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:44:07.546169018 +0000 UTC m=+35.630706942 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.546261 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.555598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.555690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.555713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.555746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.555769 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.570988 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.592241 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.592360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.592241 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.592435 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.592603 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.592740 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.603001 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.647722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.647817 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.647881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.647967 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.647999 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648032 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648155 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:07.648118677 +0000 UTC m=+35.732656591 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648182 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:07.648169879 +0000 UTC m=+35.732707793 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648253 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648276 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648298 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648312 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648353 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:07.648333873 +0000 UTC m=+35.732871787 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648358 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648393 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:59 crc kubenswrapper[4678]: E1013 12:43:59.648495 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:07.648462436 +0000 UTC m=+35.733000480 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.652942 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.676028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.676115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.676134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.676164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.676188 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.680130 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.779036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.779136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.779155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.779183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.779205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.799168 4678 generic.go:334] "Generic (PLEG): container finished" podID="2dee984d-00a7-4d89-a39f-1cfaac6bb2f6" containerID="3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6" exitCode=0 Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.799286 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerDied","Data":"3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.824255 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.849746 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.873538 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.883231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.883260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.883272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.883290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.883301 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.896640 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.924190 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.941662 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.966683 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:43:59Z is after 2025-08-24T17:21:41Z" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.986815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.986856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.986870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.986889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:43:59 crc kubenswrapper[4678]: I1013 12:43:59.986902 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:43:59Z","lastTransitionTime":"2025-10-13T12:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.011292 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.038405 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.080880 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.089562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.089598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.089610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.089628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.089641 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.123440 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.165240 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.192560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.192600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.192610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.192625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.192633 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.202395 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.243858 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.282032 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.295217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.295263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.295273 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.295289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.295303 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.397611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.397661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.397677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.397701 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.397720 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.500694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.500755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.500773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.500797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.500825 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.604422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.604489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.604506 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.604532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.604553 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.707656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.707716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.707733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.707759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.707779 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.809605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.809660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.809677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.809703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.809721 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.810438 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.811155 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.811333 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.811355 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.818564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" event={"ID":"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6","Type":"ContainerStarted","Data":"fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.833981 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.894285 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.894386 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.895362 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.906803 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.912071 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.912172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.912186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.912202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.912236 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:00Z","lastTransitionTime":"2025-10-13T12:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.935513 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.950635 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.966958 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:00 crc kubenswrapper[4678]: I1013 12:44:00.985134 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.001452 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.020607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.020664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.020677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.020696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.020709 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.027908 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.044329 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.060543 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.087588 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.097996 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.112880 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.123152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.123220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.123241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.123265 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.123282 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.126866 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.142898 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.162166 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.178210 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.197556 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.213429 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.226387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.226448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.226473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.226508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.226532 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.229079 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.247631 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.265995 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.284124 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.304963 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.329623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.329682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.329694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.329710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.329725 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.336559 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.361471 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.407423 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.432506 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.432808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.432965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.433196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.433407 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.454027 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.480547 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:01Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.536966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.537028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.537045 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.537101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.537121 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.591578 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.591595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.591641 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:01 crc kubenswrapper[4678]: E1013 12:44:01.591740 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:01 crc kubenswrapper[4678]: E1013 12:44:01.591910 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:01 crc kubenswrapper[4678]: E1013 12:44:01.592154 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.640715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.640772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.640789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.640815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.640832 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.743362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.743428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.743448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.743472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.743489 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.847034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.847123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.847141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.847168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.847189 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.949499 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.949561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.949579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.949602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:01 crc kubenswrapper[4678]: I1013 12:44:01.949619 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:01Z","lastTransitionTime":"2025-10-13T12:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.053176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.053251 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.053278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.053310 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.053333 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.156686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.156731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.156748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.156769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.156781 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.259433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.259480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.259494 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.259514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.259529 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.361980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.362116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.362135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.362157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.362175 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.465160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.465238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.465260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.465290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.465312 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.568285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.568332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.568344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.568362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.568375 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.615206 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.642959 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.670299 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.671583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.671645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.671658 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.671676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.671709 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.691412 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.710880 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.734662 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.748450 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.759950 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.771700 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.773886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.773935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.773943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.773956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.773966 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.782166 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.795398 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.808017 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.822632 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.840174 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.853622 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.876737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.876778 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.876790 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.876808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.876820 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.979478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.980081 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.980224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.980392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:02 crc kubenswrapper[4678]: I1013 12:44:02.980485 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:02Z","lastTransitionTime":"2025-10-13T12:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.083486 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.083553 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.083577 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.083608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.083632 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.186400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.186447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.186463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.186487 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.186519 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.289886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.289948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.289964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.289988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.290007 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.393358 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.393414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.393431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.393454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.393471 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.496848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.496910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.496930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.496956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.496975 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.592377 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.592426 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.592378 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:03 crc kubenswrapper[4678]: E1013 12:44:03.592596 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:03 crc kubenswrapper[4678]: E1013 12:44:03.592721 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:03 crc kubenswrapper[4678]: E1013 12:44:03.592800 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.600009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.600168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.600186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.600207 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.600223 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.703300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.703353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.703367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.703385 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.703401 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.806285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.806331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.806342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.806379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.806396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.833606 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/0.log" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.837205 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633" exitCode=1 Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.837277 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.838651 4678 scope.go:117] "RemoveContainer" containerID="bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.864575 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.885678 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.907835 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.911044 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.911331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.911469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.911609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.911739 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:03Z","lastTransitionTime":"2025-10-13T12:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.930143 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.947632 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:03 crc kubenswrapper[4678]: I1013 12:44:03.982352 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:03Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.005138 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.015192 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.015247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.015267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.015290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.015307 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.028026 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.051127 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.071363 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.095430 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.107991 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.117391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.117469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.117493 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.117524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.117547 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.126465 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.148043 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.160597 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.219713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.219780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.219797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.219822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.219838 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.325335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.325391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.325412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.325450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.325469 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.428215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.428265 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.428278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.428297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.428309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.530364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.530400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.530410 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.530426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.530435 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.633469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.633527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.633545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.633567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.633584 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.735913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.735952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.735961 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.735977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.735987 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.839620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.839662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.839673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.839691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.839703 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.842231 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/1.log" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.842934 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/0.log" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.845390 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749" exitCode=1 Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.845425 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.845456 4678 scope.go:117] "RemoveContainer" containerID="bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.846479 4678 scope.go:117] "RemoveContainer" containerID="def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749" Oct 13 12:44:04 crc kubenswrapper[4678]: E1013 12:44:04.846731 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.865671 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.884928 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.902122 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.917478 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.935396 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.941754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.941781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.941789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.941802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.941812 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:04Z","lastTransitionTime":"2025-10-13T12:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.949819 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.964549 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.982041 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:04 crc kubenswrapper[4678]: I1013 12:44:04.998330 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.011998 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.026275 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.036468 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.044140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.044175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.044183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.044196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.044204 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.048260 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.062720 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.074177 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:05Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.147225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.147290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.147308 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.147333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.147351 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.252949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.253027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.253046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.253103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.253131 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.356408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.356473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.356488 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.356512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.356528 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.459292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.459467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.459495 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.459528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.459549 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.563020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.563136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.563212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.563261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.563283 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.591500 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.591540 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.591597 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:05 crc kubenswrapper[4678]: E1013 12:44:05.591719 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:05 crc kubenswrapper[4678]: E1013 12:44:05.591856 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:05 crc kubenswrapper[4678]: E1013 12:44:05.591981 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.666771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.666840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.666857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.666881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.666898 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.770284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.770353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.770371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.770395 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.770411 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.851321 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/1.log" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.873119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.873187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.873214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.873243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.873264 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.976566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.976621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.976638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.976662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:05 crc kubenswrapper[4678]: I1013 12:44:05.976681 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:05Z","lastTransitionTime":"2025-10-13T12:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.079345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.079408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.079425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.079450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.079470 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.182660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.182705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.182721 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.182744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.182761 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.285836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.285893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.285909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.285935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.285953 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.367464 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k"] Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.368149 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.370358 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.370744 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.388941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.389324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.389342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.389365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.389384 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.395197 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.412602 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.424594 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.446808 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.459910 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.474594 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.474650 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.474695 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.474734 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-475dd\" (UniqueName: \"kubernetes.io/projected/c1c320f8-4efb-44fd-8ad0-fa27faea2811-kube-api-access-475dd\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.476446 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.490766 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.492323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.492384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.492402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.492428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.492446 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.505968 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.522887 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.539632 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.563206 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.576302 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-475dd\" (UniqueName: \"kubernetes.io/projected/c1c320f8-4efb-44fd-8ad0-fa27faea2811-kube-api-access-475dd\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.576426 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.576467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.576521 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.577436 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-env-overrides\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.577808 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.585202 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c1c320f8-4efb-44fd-8ad0-fa27faea2811-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598455 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.598341 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.610961 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-475dd\" (UniqueName: \"kubernetes.io/projected/c1c320f8-4efb-44fd-8ad0-fa27faea2811-kube-api-access-475dd\") pod \"ovnkube-control-plane-749d76644c-n5l5k\" (UID: \"c1c320f8-4efb-44fd-8ad0-fa27faea2811\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.617682 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.635227 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.654415 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.668441 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:06Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.687102 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.701415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.701453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.701466 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.701488 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.701503 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: W1013 12:44:06.706878 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1c320f8_4efb_44fd_8ad0_fa27faea2811.slice/crio-16149714cee803bbf4ca83471c859f20573cb96a92a4e949ef86c2072b63e015 WatchSource:0}: Error finding container 16149714cee803bbf4ca83471c859f20573cb96a92a4e949ef86c2072b63e015: Status 404 returned error can't find the container with id 16149714cee803bbf4ca83471c859f20573cb96a92a4e949ef86c2072b63e015 Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.803701 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.803782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.803803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.803828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.803849 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.859711 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" event={"ID":"c1c320f8-4efb-44fd-8ad0-fa27faea2811","Type":"ContainerStarted","Data":"16149714cee803bbf4ca83471c859f20573cb96a92a4e949ef86c2072b63e015"} Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.906043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.906138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.906164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.906201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:06 crc kubenswrapper[4678]: I1013 12:44:06.906227 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:06Z","lastTransitionTime":"2025-10-13T12:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.009546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.009590 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.009602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.009619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.009632 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.112926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.112987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.113007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.113032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.113079 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.215724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.215788 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.215810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.215835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.215887 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.319126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.319184 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.319204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.319227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.319245 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.421974 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.422037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.422095 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.422122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.422139 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.525440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.525509 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.525525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.525554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.525571 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.587548 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.587755 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.587715098 +0000 UTC m=+51.672253022 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.591691 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.591799 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.591867 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.591942 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.592166 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.592321 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.629183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.629236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.629253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.629278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.629298 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.689492 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.689579 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.689631 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.689672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689702 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689749 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689768 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689773 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689853 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.689828342 +0000 UTC m=+51.774366266 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689879 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.689867503 +0000 UTC m=+51.774405427 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689888 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689886 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689916 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.689949 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.690019 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.689991246 +0000 UTC m=+51.774529170 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.690048 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.690035617 +0000 UTC m=+51.774573541 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.732855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.732923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.732939 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.732964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.732982 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.836007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.836188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.836214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.836242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.836262 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.866827 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" event={"ID":"c1c320f8-4efb-44fd-8ad0-fa27faea2811","Type":"ContainerStarted","Data":"3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.866895 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" event={"ID":"c1c320f8-4efb-44fd-8ad0-fa27faea2811","Type":"ContainerStarted","Data":"b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.891041 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:07Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.904431 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hbs76"] Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.905280 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:07 crc kubenswrapper[4678]: E1013 12:44:07.905400 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.911255 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:07Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.931584 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:07Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.939128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.939191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.939210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.939236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.939254 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:07Z","lastTransitionTime":"2025-10-13T12:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.952444 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:07Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.968903 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:07Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.997623 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47mr2\" (UniqueName: \"kubernetes.io/projected/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-kube-api-access-47mr2\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:07 crc kubenswrapper[4678]: I1013 12:44:07.998233 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.005785 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.027453 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.042430 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.042508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.042532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.042564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.042587 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.047827 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.066749 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.084789 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.100722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.100825 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47mr2\" (UniqueName: \"kubernetes.io/projected/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-kube-api-access-47mr2\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.100855 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.100948 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:08.600920743 +0000 UTC m=+36.685458667 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.106534 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.123860 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.131723 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47mr2\" (UniqueName: \"kubernetes.io/projected/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-kube-api-access-47mr2\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.145541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.145591 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.145608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.145636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.145653 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.148932 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.179138 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.196766 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.218368 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.238826 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.248299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.248401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.248421 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.248444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.248461 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.263446 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.295676 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.313859 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.331991 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.348515 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.352645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.352687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.352699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.352715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.352727 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.363453 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.377160 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.396514 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.412639 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.423235 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.450968 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.454947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.454982 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.454995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.455012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.455023 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.467563 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.481410 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.497836 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.509397 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.523289 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.557429 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.557732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.557872 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.558034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.558206 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.606019 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.606235 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.606294 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:09.6062787 +0000 UTC m=+37.690816594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.661176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.661225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.661237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.661255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.661268 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.762213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.762558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.762757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.763011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.763207 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.786566 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.790999 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.791044 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.791080 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.791102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.791116 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.805845 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.810111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.810157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.810169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.810185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.810199 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.826446 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.830533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.830596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.830621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.830653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.830676 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.844535 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.848924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.848971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.848983 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.849002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.849013 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.865595 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:08Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:08 crc kubenswrapper[4678]: E1013 12:44:08.865736 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.868148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.868232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.868250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.868277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.868295 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.971260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.971340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.971367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.971398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:08 crc kubenswrapper[4678]: I1013 12:44:08.971420 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:08Z","lastTransitionTime":"2025-10-13T12:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.074755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.074817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.074834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.074857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.074873 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.177622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.177666 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.177679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.177724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.177737 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.280010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.280119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.280137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.280162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.280179 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.383038 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.383128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.383141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.383160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.383173 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.486166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.486216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.486229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.486246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.486261 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.588522 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.588594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.588614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.588639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.588656 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.592170 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.592256 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.592201 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.592407 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.592217 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.592604 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.592706 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.592760 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.617767 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.617992 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:09 crc kubenswrapper[4678]: E1013 12:44:09.618089 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:11.618036508 +0000 UTC m=+39.702574422 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.691107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.691162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.691174 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.691194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.691205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.794534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.794589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.794608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.794632 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.794648 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.897619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.897676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.897693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.897716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:09 crc kubenswrapper[4678]: I1013 12:44:09.897739 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:09Z","lastTransitionTime":"2025-10-13T12:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.000388 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.000483 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.000504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.000527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.000543 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.103506 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.103571 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.103595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.103621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.103638 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.206889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.206953 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.206969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.206993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.207010 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.310118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.310183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.310201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.310225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.310243 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.413048 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.413118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.413126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.413139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.413147 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.516201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.516277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.516301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.516332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.516355 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.619498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.619576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.619597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.619626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.619649 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.722178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.722247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.722264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.722292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.722309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.824668 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.824754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.824773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.825303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.825367 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.928567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.928623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.928634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.928668 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:10 crc kubenswrapper[4678]: I1013 12:44:10.928683 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:10Z","lastTransitionTime":"2025-10-13T12:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.031908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.031976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.031993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.032018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.032034 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.135237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.135305 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.135323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.135346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.135364 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.238928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.239027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.239092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.239124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.239142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.342024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.342099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.342115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.342137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.342153 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.445380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.445442 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.445459 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.445482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.445503 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.548095 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.548168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.548193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.548228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.548251 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.591375 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.591459 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.591546 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.591711 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.591760 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.591903 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.592047 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.592328 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.639734 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.639986 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:11 crc kubenswrapper[4678]: E1013 12:44:11.640120 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:15.64009148 +0000 UTC m=+43.724629404 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.651724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.651764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.651781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.651804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.651822 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.755155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.755715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.755829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.755862 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.755885 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.858980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.859166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.859230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.859256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.859272 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.961550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.961613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.961633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.961665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:11 crc kubenswrapper[4678]: I1013 12:44:11.961683 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:11Z","lastTransitionTime":"2025-10-13T12:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.064975 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.065041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.065091 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.065126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.065144 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.168205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.168267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.168285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.168311 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.168329 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.271454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.271535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.271559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.271595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.271613 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.374880 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.374949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.374966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.374990 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.375012 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.477561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.477625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.477642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.477665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.477682 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.580804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.580876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.580901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.580933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.580956 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.618044 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.638391 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.654302 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.675351 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.684027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.684179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.684277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.684315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.684342 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.700014 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.719741 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.736197 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.754926 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.785368 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.786702 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.786774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.786796 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.786821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.786839 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.801039 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.822215 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.848302 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb6ccefd2e73991d176428d378a57fa9df0c3a41d0478af30acbfd9508b13633\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:02Z\\\",\\\"message\\\":\\\"y (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.846977 5978 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847113 5978 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1013 12:44:02.847205 5978 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1013 12:44:02.847505 5978 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1013 12:44:02.847737 5978 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1013 12:44:02.847790 5978 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1013 12:44:02.847797 5978 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1013 12:44:02.847813 5978 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1013 12:44:02.847818 5978 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1013 12:44:02.847838 5978 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1013 12:44:02.848933 5978 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.860710 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.875292 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.889486 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.890122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.890180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.891301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.891332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.891352 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.905445 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.923765 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.994746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.994779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.994789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.994806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:12 crc kubenswrapper[4678]: I1013 12:44:12.994817 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:12Z","lastTransitionTime":"2025-10-13T12:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.096992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.097105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.097130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.097158 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.097176 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.199967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.200025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.200041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.200099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.200120 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.303196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.304156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.304312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.304456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.304591 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.407625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.407694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.407718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.407744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.407762 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.510503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.510555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.510570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.510590 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.510605 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.591503 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:13 crc kubenswrapper[4678]: E1013 12:44:13.591603 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.591692 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.591764 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.591789 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:13 crc kubenswrapper[4678]: E1013 12:44:13.591939 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:13 crc kubenswrapper[4678]: E1013 12:44:13.592041 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:13 crc kubenswrapper[4678]: E1013 12:44:13.592178 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.613366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.613433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.613460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.613488 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.613505 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.715762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.716134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.716287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.716419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.716532 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.818684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.818746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.818763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.818785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.818803 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.922022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.922113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.922130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.922154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:13 crc kubenswrapper[4678]: I1013 12:44:13.922171 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:13Z","lastTransitionTime":"2025-10-13T12:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.025280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.025336 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.025355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.025379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.025396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.128176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.128243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.128267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.128298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.128322 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.231223 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.231526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.231657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.231793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.231911 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.335777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.335850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.335874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.335902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.335922 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.439114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.439203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.439222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.439245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.439262 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.541866 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.541900 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.541911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.541927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.541938 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.644527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.644589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.644608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.644633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.644650 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.748024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.748140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.748164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.748194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.748218 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.851702 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.851819 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.851887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.851922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.852009 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.954763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.954835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.954858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.954889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:14 crc kubenswrapper[4678]: I1013 12:44:14.954911 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:14Z","lastTransitionTime":"2025-10-13T12:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.058223 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.058282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.058299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.058323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.058344 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.161480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.161550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.161573 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.161604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.161629 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.264920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.264977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.264993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.265016 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.265033 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.367389 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.367437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.367447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.367463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.367474 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.469882 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.469936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.469951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.469974 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.469996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.573824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.573884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.573902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.573942 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.573959 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.592286 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.592322 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.592401 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.592424 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.592589 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.593332 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.593564 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.593772 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.677292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.677348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.677368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.677393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.677411 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.687007 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.687308 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:15 crc kubenswrapper[4678]: E1013 12:44:15.687397 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:23.687376424 +0000 UTC m=+51.771914338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.781049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.781501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.781660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.781816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.781960 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.885302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.885348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.885387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.885407 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.885418 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.988292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.988358 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.988380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.988426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:15 crc kubenswrapper[4678]: I1013 12:44:15.988455 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:15Z","lastTransitionTime":"2025-10-13T12:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.091506 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.091556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.091574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.091596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.091614 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.194160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.194218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.194237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.194266 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.194283 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.297180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.297263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.297287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.297322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.297348 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.399493 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.399557 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.399582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.399611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.399633 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.507985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.508042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.508089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.508114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.508133 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.611221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.611284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.611302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.611328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.611346 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.714299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.714371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.714388 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.714414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.714432 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.816693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.816741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.816752 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.816769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.816783 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.919484 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.919563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.919585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.919609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:16 crc kubenswrapper[4678]: I1013 12:44:16.919630 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:16Z","lastTransitionTime":"2025-10-13T12:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.022600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.022650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.022672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.022698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.022718 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.125699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.125797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.125821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.125849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.125870 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.228741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.228811 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.228830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.228854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.228871 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.331880 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.331918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.331926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.331940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.331950 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.434620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.434661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.434672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.434706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.434718 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.538324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.538396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.538413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.538438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.538457 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.591431 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.591455 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.591512 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.591574 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:17 crc kubenswrapper[4678]: E1013 12:44:17.591775 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:17 crc kubenswrapper[4678]: E1013 12:44:17.592198 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:17 crc kubenswrapper[4678]: E1013 12:44:17.592162 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:17 crc kubenswrapper[4678]: E1013 12:44:17.592320 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.641881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.641979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.642000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.642026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.642045 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.745208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.745259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.745279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.745306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.745323 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.848433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.848499 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.848519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.848545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.848565 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.951179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.951262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.951280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.951313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:17 crc kubenswrapper[4678]: I1013 12:44:17.951331 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:17Z","lastTransitionTime":"2025-10-13T12:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.054791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.054857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.054874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.054898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.054916 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.157793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.157859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.157876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.157907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.157929 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.260759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.260830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.260847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.260875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.260893 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.364132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.364189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.364206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.364228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.364246 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.467747 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.467826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.467853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.467883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.467904 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.570723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.570779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.570836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.570860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.570878 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.593286 4678 scope.go:117] "RemoveContainer" containerID="def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.618169 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.640492 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.659216 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.674959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.675041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.675100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.675135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.675153 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.682142 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.714950 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.736471 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.760799 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.780412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.780473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.780490 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.780519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.780541 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.793544 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.812529 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.832756 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.849986 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.868212 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.883863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.883923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.883947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.883973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.883992 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.891250 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.910603 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/1.log" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.911876 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.916832 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.917529 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.933453 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.949508 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.969905 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.987128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.987200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.987225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.987259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.987283 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:18Z","lastTransitionTime":"2025-10-13T12:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:18 crc kubenswrapper[4678]: I1013 12:44:18.990288 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:18Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.006701 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.029392 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.050013 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.070682 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.090462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.090542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.090566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.090602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.090624 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.093158 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.101224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.101272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.101286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.101306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.101321 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.117771 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.123386 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.127616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.127679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.127698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.127726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.127743 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.150491 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.156099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.156154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.156173 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.156197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.156214 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.161133 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.181866 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.186739 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.187189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.187276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.187297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.187321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.187342 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.214092 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.223142 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.225209 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.225247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.225259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.225280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.225292 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.251158 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.251352 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253421 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253460 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.253659 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.269983 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.283626 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.294828 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.308402 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.324111 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.337241 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.356136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.356168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.356177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.356192 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.356202 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.458836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.458879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.458891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.458917 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.458931 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.561224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.561273 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.561288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.561309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.561321 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.591876 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.591927 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.591956 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.592043 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.592120 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.592318 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.592470 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.592495 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.663466 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.663513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.663525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.663541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.663578 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.766661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.766727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.766750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.766779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.766802 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.869806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.869853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.869865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.869882 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.869894 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.923484 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/2.log" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.924697 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/1.log" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.928558 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" exitCode=1 Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.928616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.928685 4678 scope.go:117] "RemoveContainer" containerID="def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.929724 4678 scope.go:117] "RemoveContainer" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" Oct 13 12:44:19 crc kubenswrapper[4678]: E1013 12:44:19.930029 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.954841 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.972950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.973012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.973031 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.973089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.973109 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:19Z","lastTransitionTime":"2025-10-13T12:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.976612 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:19 crc kubenswrapper[4678]: I1013 12:44:19.997676 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.019493 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.037221 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.072549 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.075473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.075534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.075555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.075579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.075596 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.093714 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.113799 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.131406 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.147843 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.169948 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.178504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.178574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.178601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.178634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.178664 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.187270 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.205538 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.228520 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.260975 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://def3d207270e59baf29291fe9d8f99a9b34daa9628633ed9b8838ddf24388749\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:04Z\\\",\\\"message\\\":\\\"47\\\\\\\" in cache\\\\nI1013 12:44:04.803403 6116 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-925jf after 0 failed attempt(s)\\\\nI1013 12:44:04.804963 6116 services_controller.go:453] Built service openshift-machine-config-operator/machine-config-daemon template LB for network=default: []services.LB{}\\\\nI1013 12:44:04.804971 6116 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-925jf\\\\nF1013 12:44:04.803724 6116 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:04Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:04.804977 6116 services_controller.go:454] \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.277241 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.281874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.281915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.281931 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.281952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.281968 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.296283 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.385136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.385200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.385221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.385247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.385268 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.494342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.494422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.494442 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.494477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.494498 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.597478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.597560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.597587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.597618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.597641 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.701021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.701172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.701199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.701227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.701251 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.804884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.804968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.804993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.805026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.805081 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.908820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.908896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.908914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.908939 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.908961 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:20Z","lastTransitionTime":"2025-10-13T12:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.935124 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/2.log" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.940500 4678 scope.go:117] "RemoveContainer" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" Oct 13 12:44:20 crc kubenswrapper[4678]: E1013 12:44:20.940771 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.958846 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:20 crc kubenswrapper[4678]: I1013 12:44:20.981830 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:20Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.009752 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.012140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.012214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.012240 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.012272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.012296 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.024396 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.041353 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.055887 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.076041 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.093963 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.111140 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.115941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.116029 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.116078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.116106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.116124 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.129864 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.145166 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.175711 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.194660 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.208091 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.218833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.218915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.218938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.218972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.218996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.222641 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.237102 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.255528 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:21Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.322390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.322452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.322475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.322503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.322527 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.425911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.425975 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.425991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.426015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.426036 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.529526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.529587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.529604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.529629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.529646 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.592247 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.592347 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:21 crc kubenswrapper[4678]: E1013 12:44:21.592393 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.592269 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.592514 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:21 crc kubenswrapper[4678]: E1013 12:44:21.592712 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:21 crc kubenswrapper[4678]: E1013 12:44:21.592825 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:21 crc kubenswrapper[4678]: E1013 12:44:21.593012 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.633198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.633250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.633264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.633283 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.633294 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.736645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.736748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.736768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.736795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.736824 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.840205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.840262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.840279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.840303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.840321 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.943366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.943428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.943448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.943477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:21 crc kubenswrapper[4678]: I1013 12:44:21.943502 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:21Z","lastTransitionTime":"2025-10-13T12:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.046123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.046180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.046198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.046221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.046241 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.149363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.149424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.149441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.149467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.149486 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.252791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.252868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.252881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.252899 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.252912 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.356262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.356324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.356346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.356375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.356396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.459317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.459364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.459375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.459391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.459402 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.562024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.562099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.562110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.562126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.562138 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.608370 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.632453 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.650803 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.664594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.664674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.664691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.664717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.664734 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.684738 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.705978 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.726340 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.741293 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.757203 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.768302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.768371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.768397 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.768431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.768455 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.777875 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.795660 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.821592 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.853451 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.872358 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.873741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.873800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.873821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.873908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.873932 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.890024 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.907019 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.928304 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.945701 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:22Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.977435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.977512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.977537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.977575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:22 crc kubenswrapper[4678]: I1013 12:44:22.977602 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:22Z","lastTransitionTime":"2025-10-13T12:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.080577 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.080613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.080626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.080645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.080658 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.183748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.183804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.183824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.183848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.183866 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.287444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.287510 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.287527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.287552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.287570 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.390725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.390853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.390921 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.390991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.391017 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.494607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.494670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.494689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.494712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.494732 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.591455 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.591495 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.591533 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.591498 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.591640 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.591778 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.591868 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.591983 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.597981 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.598198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.598317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.598446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.598630 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.675090 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.675281 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:44:55.675249593 +0000 UTC m=+83.759787507 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.701794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.701850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.701899 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.701922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.701939 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.776246 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.776467 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.776743 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:55.77670854 +0000 UTC m=+83.861246464 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.776600 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.776911 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.777004 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.777124 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777130 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777253 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777287 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777299 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:55.777276844 +0000 UTC m=+83.861814768 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777311 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777374 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:55.777355406 +0000 UTC m=+83.861893400 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777509 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777563 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777581 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777658 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:44:55.777643824 +0000 UTC m=+83.862181748 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.777969 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: E1013 12:44:23.778202 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:44:39.778176147 +0000 UTC m=+67.862714071 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.804854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.804907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.804923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.804948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.804968 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.907441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.907739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.907886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.908031 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:23 crc kubenswrapper[4678]: I1013 12:44:23.908264 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:23Z","lastTransitionTime":"2025-10-13T12:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.011672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.011744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.011771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.011803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.011847 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.115270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.115344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.115369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.115400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.115421 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.218802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.218883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.218902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.218928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.218945 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.322120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.322203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.322230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.322303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.322331 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.425649 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.425711 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.425729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.425755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.425773 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.528930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.529318 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.529461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.529603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.529759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.632447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.632515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.632534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.632559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.632579 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.736130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.736988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.737168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.737313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.737434 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.840567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.840636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.840654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.840677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.840695 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.943920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.943981 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.944013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.944036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:24 crc kubenswrapper[4678]: I1013 12:44:24.944086 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:24Z","lastTransitionTime":"2025-10-13T12:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.049008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.049164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.049184 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.049238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.049259 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.054291 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.068436 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.068823 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.085628 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.108986 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.141273 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.152329 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.152366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.152378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.152396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.152413 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.159333 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.176628 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.199395 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.219509 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.240820 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.255254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.255315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.255333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.255357 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.255375 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.261124 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.278740 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.310827 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.328370 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.346216 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.357800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.357870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.357897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.357928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.357981 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.364257 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.380467 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.403079 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:25Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.460722 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.460783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.460800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.460827 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.460851 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.564202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.564282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.564309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.564338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.564360 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.591748 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.591802 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.591831 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:25 crc kubenswrapper[4678]: E1013 12:44:25.591868 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:25 crc kubenswrapper[4678]: E1013 12:44:25.591991 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.592086 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:25 crc kubenswrapper[4678]: E1013 12:44:25.592258 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:25 crc kubenswrapper[4678]: E1013 12:44:25.592449 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.667283 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.667363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.667391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.667427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.667450 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.770748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.770820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.771022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.771108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.771133 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.874865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.874928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.874949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.874974 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.874991 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.977887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.977971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.977991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.978018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:25 crc kubenswrapper[4678]: I1013 12:44:25.978035 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:25Z","lastTransitionTime":"2025-10-13T12:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.081039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.081127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.081146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.081171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.081194 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.183520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.183607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.183625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.183645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.183659 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.286417 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.286478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.286496 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.286523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.286540 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.389663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.389756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.389773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.389795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.389813 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.492737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.492806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.492824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.492848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.492883 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.601941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.602028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.602096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.602136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.602160 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.704766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.704837 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.704863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.704891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.704914 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.808126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.808189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.808206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.808233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.808251 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.911723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.911791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.911809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.911835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:26 crc kubenswrapper[4678]: I1013 12:44:26.911853 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:26Z","lastTransitionTime":"2025-10-13T12:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.014011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.014100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.014120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.014144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.014161 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.117521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.117622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.117659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.119101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.119138 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.221829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.221890 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.221907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.221929 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.221946 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.325464 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.325563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.325583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.325610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.325629 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.428558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.428621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.428640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.428665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.428684 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.531549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.531626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.531645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.531674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.531696 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.591904 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.591992 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.592025 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.591932 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:27 crc kubenswrapper[4678]: E1013 12:44:27.592183 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:27 crc kubenswrapper[4678]: E1013 12:44:27.592299 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:27 crc kubenswrapper[4678]: E1013 12:44:27.592456 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:27 crc kubenswrapper[4678]: E1013 12:44:27.592662 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.636802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.636935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.637009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.637045 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.637128 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.739937 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.740024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.740049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.740121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.740144 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.844012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.844840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.844997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.845187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.845363 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.948092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.948146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.948166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.948191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:27 crc kubenswrapper[4678]: I1013 12:44:27.948208 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:27Z","lastTransitionTime":"2025-10-13T12:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.050940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.051005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.051023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.051048 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.051094 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.153832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.153870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.153879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.153893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.153903 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.256252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.256307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.256323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.256345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.256360 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.358138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.358168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.358176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.358188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.358196 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.460963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.461001 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.461008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.461020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.461029 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.563474 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.563532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.563554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.563584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.563621 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.666593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.666670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.666688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.666710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.666729 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.769555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.769632 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.769653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.769677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.769697 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.872564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.872621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.872639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.872662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.872679 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.975589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.975652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.975673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.975699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:28 crc kubenswrapper[4678]: I1013 12:44:28.975722 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:28Z","lastTransitionTime":"2025-10-13T12:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.078821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.078888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.078911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.078940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.078958 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.182249 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.182351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.182368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.182393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.182411 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.285503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.285566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.285583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.285608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.285626 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.379290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.379368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.379391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.379423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.379445 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.401819 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:29Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.405776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.405834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.405861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.405894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.405918 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.425693 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:29Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.429650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.429682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.429693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.429710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.429722 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.443233 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:29Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.454028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.454091 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.454102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.454117 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.454129 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.524996 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:29Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.529077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.529113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.529122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.529136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.529146 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.543429 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:29Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.543536 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.544973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.544994 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.545003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.545013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.545022 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.591988 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.592043 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.592083 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.592071 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.592173 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.592306 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.592416 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:29 crc kubenswrapper[4678]: E1013 12:44:29.592528 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.647271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.647320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.647334 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.647351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.647363 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.749544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.749604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.749625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.749652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.749668 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.852489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.852546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.852563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.852585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.852605 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.955292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.955330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.955346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.955366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:29 crc kubenswrapper[4678]: I1013 12:44:29.955382 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:29Z","lastTransitionTime":"2025-10-13T12:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.058507 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.058552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.058569 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.058593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.058609 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.161219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.161280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.161297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.161324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.161342 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.264554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.264625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.264645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.264674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.264696 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.367900 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.367960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.367978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.368003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.368023 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.471333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.471402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.471429 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.471462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.471483 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.574615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.574692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.574724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.574757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.574777 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.677441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.677510 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.677526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.677549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.677565 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.779791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.779855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.779874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.779898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.779916 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.882884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.882954 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.882976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.883009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.883030 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.985809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.986855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.987100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.987289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:30 crc kubenswrapper[4678]: I1013 12:44:30.987434 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:30Z","lastTransitionTime":"2025-10-13T12:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.090136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.090203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.090221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.090245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.090262 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.193734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.193792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.193810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.193835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.193852 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.296894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.296956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.296973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.296999 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.297016 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.399733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.399784 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.399801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.399820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.399836 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.502532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.502607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.502630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.502660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.502683 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.592269 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.592366 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.592288 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.592280 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:31 crc kubenswrapper[4678]: E1013 12:44:31.592462 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:31 crc kubenswrapper[4678]: E1013 12:44:31.592640 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:31 crc kubenswrapper[4678]: E1013 12:44:31.592730 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:31 crc kubenswrapper[4678]: E1013 12:44:31.592873 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.606303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.606355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.606372 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.606396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.606415 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.709685 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.709750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.709768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.709794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.709811 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.813396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.813472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.813495 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.813525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.813547 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.916808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.916867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.916885 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.916909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:31 crc kubenswrapper[4678]: I1013 12:44:31.916926 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:31Z","lastTransitionTime":"2025-10-13T12:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.019467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.019567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.019585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.019616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.019635 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.122532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.122586 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.122602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.122626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.122642 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.225350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.225414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.225435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.225460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.225479 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.329162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.329220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.329237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.329260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.329277 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.432675 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.432738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.432756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.432782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.432801 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.535567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.535626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.535646 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.535670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.535687 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.627351 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.639817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.639890 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.639908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.639932 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.639948 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.649249 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.668948 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.689282 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.711596 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.732046 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.742548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.742638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.742661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.742687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.742738 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.750828 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.766120 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.791494 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.817626 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.835572 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.846133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.846518 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.846719 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.846900 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.847107 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.855746 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.874711 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.895737 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.916278 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.936803 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.950456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.950558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.950576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.950601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.950618 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:32Z","lastTransitionTime":"2025-10-13T12:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.956319 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:32 crc kubenswrapper[4678]: I1013 12:44:32.972733 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:32Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.053725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.054803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.055107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.055394 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.055624 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.159528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.159902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.160147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.160346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.160517 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.265494 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.266229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.266471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.266679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.266848 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.370070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.370098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.370105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.370119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.370127 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.479650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.479690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.479699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.479713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.479723 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.583470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.583897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.584087 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.584269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.584398 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.592124 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.592198 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.592236 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:33 crc kubenswrapper[4678]: E1013 12:44:33.592687 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:33 crc kubenswrapper[4678]: E1013 12:44:33.592508 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.592273 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:33 crc kubenswrapper[4678]: E1013 12:44:33.592838 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:33 crc kubenswrapper[4678]: E1013 12:44:33.592946 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.687774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.687839 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.687859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.687884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.687900 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.791010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.791135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.791159 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.791189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.791212 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.894727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.894795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.894818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.894849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.894870 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.997795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.998039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.998099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.998124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:33 crc kubenswrapper[4678]: I1013 12:44:33.998142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:33Z","lastTransitionTime":"2025-10-13T12:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.101575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.101652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.101679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.101710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.101733 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.204737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.204809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.204835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.204865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.204887 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.307386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.307449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.307469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.307495 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.307512 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.410367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.410406 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.410417 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.410434 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.410445 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.513016 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.513093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.513103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.513120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.513130 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.592701 4678 scope.go:117] "RemoveContainer" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" Oct 13 12:44:34 crc kubenswrapper[4678]: E1013 12:44:34.593107 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.614556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.614597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.614608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.614622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.614634 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.716890 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.716947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.716963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.716987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.717005 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.819706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.819746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.819758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.819775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.819787 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.921758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.921816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.921835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.921860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:34 crc kubenswrapper[4678]: I1013 12:44:34.921877 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:34Z","lastTransitionTime":"2025-10-13T12:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.025214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.025257 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.025273 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.025295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.025311 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.128225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.128281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.128300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.128323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.128340 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.231233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.231284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.231307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.231333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.231353 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.334341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.334420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.334442 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.334472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.334495 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.437446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.437501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.437518 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.437543 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.437560 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.541514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.541563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.541580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.541603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.541620 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.592333 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.592394 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.592409 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.592409 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:35 crc kubenswrapper[4678]: E1013 12:44:35.592505 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:35 crc kubenswrapper[4678]: E1013 12:44:35.592751 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:35 crc kubenswrapper[4678]: E1013 12:44:35.592888 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:35 crc kubenswrapper[4678]: E1013 12:44:35.592973 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.644569 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.644656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.644674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.644728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.644746 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.747268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.747326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.747343 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.747365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.747381 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.850022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.850080 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.850093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.850114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.850126 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.958501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.958535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.958542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.958555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:35 crc kubenswrapper[4678]: I1013 12:44:35.958563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:35Z","lastTransitionTime":"2025-10-13T12:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.068496 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.068529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.068537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.068550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.068558 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.171712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.171750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.171758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.171773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.171783 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.274139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.274214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.274224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.274237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.274247 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.376261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.376302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.376313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.376332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.376344 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.478868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.478898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.478906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.478918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.478927 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.581027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.581123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.581180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.581210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.581230 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.683795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.683852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.683868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.683891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.683911 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.786326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.786424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.786451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.786486 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.786511 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.888488 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.888556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.888578 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.888618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.888640 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.990576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.990615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.990624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.990640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:36 crc kubenswrapper[4678]: I1013 12:44:36.990651 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:36Z","lastTransitionTime":"2025-10-13T12:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.093617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.093650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.093660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.093672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.093681 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.196113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.196172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.196190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.196213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.196229 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.299250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.299309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.299328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.299353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.299372 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.401813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.401850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.401861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.401875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.401885 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.504666 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.504715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.504726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.504743 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.504756 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.591724 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.591760 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.591738 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.591847 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:37 crc kubenswrapper[4678]: E1013 12:44:37.591948 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:37 crc kubenswrapper[4678]: E1013 12:44:37.592040 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:37 crc kubenswrapper[4678]: E1013 12:44:37.592227 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:37 crc kubenswrapper[4678]: E1013 12:44:37.592282 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.607346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.607393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.607411 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.607436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.607454 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.709622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.709654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.709664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.709676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.709684 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.812594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.812624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.812641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.812660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.812678 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.915443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.915498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.915515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.915540 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:37 crc kubenswrapper[4678]: I1013 12:44:37.915558 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:37Z","lastTransitionTime":"2025-10-13T12:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.017676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.017728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.017746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.017768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.017784 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.120413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.120472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.120489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.120512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.120530 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.223548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.223596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.223608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.223624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.223635 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.325772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.325835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.325857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.325881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.325898 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.428801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.428871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.428893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.428922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.428938 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.531624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.531691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.531712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.531738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.531757 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.634710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.634744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.634753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.634766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.634775 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.737450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.737491 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.737502 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.737520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.737534 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.839491 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.839529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.839541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.839557 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.839567 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.941904 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.941944 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.941985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.941998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:38 crc kubenswrapper[4678]: I1013 12:44:38.942006 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:38Z","lastTransitionTime":"2025-10-13T12:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.044902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.044934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.044943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.044955 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.044965 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.147687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.147753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.147775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.147799 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.147817 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.250030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.250103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.250115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.250130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.250142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.352208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.352274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.352290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.352314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.352330 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.455514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.455563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.455579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.455602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.455618 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.557429 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.557499 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.557511 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.557529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.557548 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.592242 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.592276 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.592340 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.592370 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.592458 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.592527 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.592631 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.592716 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.660759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.660785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.660810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.660824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.660833 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.763307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.763350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.763360 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.763374 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.763385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.855264 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.855403 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.855453 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:45:11.855439824 +0000 UTC m=+99.939977708 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.865924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.865985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.866005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.866029 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.866048 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.930569 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.930605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.930620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.930636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.930645 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.945178 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:39Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.949384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.949425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.949436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.949450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.949461 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.968485 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:39Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.973164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.973218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.973236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.973258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.973276 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:39 crc kubenswrapper[4678]: E1013 12:44:39.987841 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:39Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.996684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.996750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.996774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.996803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:39 crc kubenswrapper[4678]: I1013 12:44:39.996822 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:39Z","lastTransitionTime":"2025-10-13T12:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: E1013 12:44:40.019887 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:40Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.024668 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.024730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.024755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.024783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.024802 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: E1013 12:44:40.045493 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:40Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:40 crc kubenswrapper[4678]: E1013 12:44:40.045742 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.047696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.047798 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.047856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.047928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.047991 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.150267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.150486 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.150544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.150622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.150691 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.252966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.253031 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.253081 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.253109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.253126 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.355369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.355427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.355444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.355468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.355484 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.458380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.458434 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.458450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.458473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.458490 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.561146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.561196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.561212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.561236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.561253 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.662988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.663023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.663035 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.663064 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.663079 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.765572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.765596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.765604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.765614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.765622 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.868080 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.868145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.868168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.868197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.868219 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.970458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.970492 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.970500 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.970517 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:40 crc kubenswrapper[4678]: I1013 12:44:40.970526 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:40Z","lastTransitionTime":"2025-10-13T12:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.015287 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/0.log" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.015315 4678 generic.go:334] "Generic (PLEG): container finished" podID="21d5191d-621a-4416-b635-e3459b3cfe8d" containerID="43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376" exitCode=1 Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.015337 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerDied","Data":"43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.015625 4678 scope.go:117] "RemoveContainer" containerID="43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.041149 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.060562 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.073419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.073633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.073641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.073654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.073662 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.080240 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.099247 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.116130 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.131318 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.141100 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.152108 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.164562 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.175199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.175221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.175239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.175252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.175261 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.180450 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.194817 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.213372 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.234009 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.250429 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.266108 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.278034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.278071 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.278079 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.278092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.278100 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.280400 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.294651 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.306889 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:41Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.380216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.380261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.380278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.380298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.380315 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.482731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.482804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.482813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.482826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.482836 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.585784 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.585889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.585906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.585931 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.585949 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.592042 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.592130 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.592169 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.592169 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:41 crc kubenswrapper[4678]: E1013 12:44:41.592253 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:41 crc kubenswrapper[4678]: E1013 12:44:41.592374 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:41 crc kubenswrapper[4678]: E1013 12:44:41.592497 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:41 crc kubenswrapper[4678]: E1013 12:44:41.592687 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.688802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.688834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.688847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.688862 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.688876 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.791163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.791204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.791218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.791235 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.791249 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.894781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.894826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.894843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.894864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.894879 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.998099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.998214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.998235 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.998258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:41 crc kubenswrapper[4678]: I1013 12:44:41.998274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:41Z","lastTransitionTime":"2025-10-13T12:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.021381 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/0.log" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.021454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerStarted","Data":"1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.034438 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.049148 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.068624 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.096501 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.102237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.102297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.102307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.102327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.102340 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.108226 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.122968 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.135448 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.154904 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.172437 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.186370 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.204040 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.205590 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.205641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.205659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.205684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.205703 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.216327 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.241548 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.257549 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.274366 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.294602 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.308138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.308167 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.308178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.308197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.308209 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.312920 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.332424 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.410695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.410726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.410737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.410751 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.410761 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.513369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.513423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.513434 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.513446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.513456 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.612686 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.615267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.615300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.615309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.615325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.615334 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.630632 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.643608 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.660421 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.674892 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.698152 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.712926 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.717309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.717363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.717382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.717408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.717427 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.726369 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.739225 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.764415 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.776491 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.787550 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.797777 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.810373 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.820145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.820186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.820197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.820213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.820222 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.822660 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.838351 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.849327 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.859413 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:42Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.923096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.923139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.923151 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.923169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:42 crc kubenswrapper[4678]: I1013 12:44:42.923182 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:42Z","lastTransitionTime":"2025-10-13T12:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.024715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.024758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.024769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.024824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.024835 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.126904 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.126957 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.126972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.126992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.127004 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.229704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.229741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.229749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.229764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.229773 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.331662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.331695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.331705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.331718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.331726 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.438196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.438257 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.438271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.438289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.438308 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.540319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.540364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.540375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.540391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.540401 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.591905 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:43 crc kubenswrapper[4678]: E1013 12:44:43.592004 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.591913 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.592030 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:43 crc kubenswrapper[4678]: E1013 12:44:43.592100 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.592148 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:43 crc kubenswrapper[4678]: E1013 12:44:43.592197 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:43 crc kubenswrapper[4678]: E1013 12:44:43.592347 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.642625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.642656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.642665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.642676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.642685 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.744911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.744943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.744952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.744966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.744975 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.848280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.848314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.848323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.848335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.848344 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.950076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.950105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.950113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.950125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:43 crc kubenswrapper[4678]: I1013 12:44:43.950133 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:43Z","lastTransitionTime":"2025-10-13T12:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.052490 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.052538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.052552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.052569 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.052582 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.155293 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.155356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.155381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.155411 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.155433 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.258180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.258254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.258279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.258307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.258361 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.360806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.360869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.360886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.360912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.360931 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.464501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.464582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.464608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.464638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.464660 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.567093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.567128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.567140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.567156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.567166 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.669419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.669489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.669514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.669548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.669572 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.772833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.772895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.772913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.772937 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.772956 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.876519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.876583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.876600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.876625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.876642 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.979904 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.979971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.979993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.980021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:44 crc kubenswrapper[4678]: I1013 12:44:44.980039 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:44Z","lastTransitionTime":"2025-10-13T12:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.082831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.082874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.082882 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.082897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.082907 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.186419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.186469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.186479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.186498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.186507 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.289176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.289247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.289265 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.289292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.289332 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.391970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.392011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.392020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.392034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.392046 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.494225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.494289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.494306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.494330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.494349 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.592281 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.592339 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.592391 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.592301 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:45 crc kubenswrapper[4678]: E1013 12:44:45.592473 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:45 crc kubenswrapper[4678]: E1013 12:44:45.592603 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:45 crc kubenswrapper[4678]: E1013 12:44:45.592745 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:45 crc kubenswrapper[4678]: E1013 12:44:45.592965 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.597550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.597588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.597600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.597616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.597629 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.700698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.700756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.700772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.700799 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.700818 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.803640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.803695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.803717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.803741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.803762 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.906433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.906533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.906552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.906578 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:45 crc kubenswrapper[4678]: I1013 12:44:45.906596 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:45Z","lastTransitionTime":"2025-10-13T12:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.009179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.009241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.009258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.009284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.009306 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.111366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.111424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.111436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.111454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.111467 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.213902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.213980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.214008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.214037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.214088 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.317289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.317326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.317335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.317349 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.317361 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.420523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.420570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.420582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.420599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.420610 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.524148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.524199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.524216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.524239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.524256 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.627269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.627338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.627360 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.627395 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.627419 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.730910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.730988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.731010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.731041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.731117 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.834017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.834087 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.834132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.834156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.834169 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.936416 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.936467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.936480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.936500 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:46 crc kubenswrapper[4678]: I1013 12:44:46.936512 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:46Z","lastTransitionTime":"2025-10-13T12:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.038992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.039099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.039123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.039154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.039177 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.142528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.142594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.142618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.142648 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.142670 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.245558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.245618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.245634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.245659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.245676 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.348196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.348256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.348272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.348296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.348313 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.451948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.452005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.452022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.452048 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.452094 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.554574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.554643 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.554662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.554689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.554710 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.591627 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.591657 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.591763 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.591811 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:47 crc kubenswrapper[4678]: E1013 12:44:47.591903 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:47 crc kubenswrapper[4678]: E1013 12:44:47.592023 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:47 crc kubenswrapper[4678]: E1013 12:44:47.592333 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:47 crc kubenswrapper[4678]: E1013 12:44:47.592615 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.657770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.657816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.657836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.657857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.657874 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.761002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.761107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.761124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.761157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.761175 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.864664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.864748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.864772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.864806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.864830 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.968356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.968414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.968432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.968456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:47 crc kubenswrapper[4678]: I1013 12:44:47.968473 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:47Z","lastTransitionTime":"2025-10-13T12:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.071894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.071969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.071993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.072046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.072132 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.174348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.174380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.174390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.174406 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.174418 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.276951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.277013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.277035 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.277092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.277117 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.379683 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.379727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.379735 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.379754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.379763 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.482392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.482465 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.482485 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.482509 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.482525 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.585189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.585260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.585282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.585308 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.585329 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.593149 4678 scope.go:117] "RemoveContainer" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.688130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.688195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.688214 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.688239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.688259 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.791266 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.791344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.791367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.791423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.791450 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.894152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.894191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.894203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.894223 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.894235 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.996771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.996818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.996833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.996854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:48 crc kubenswrapper[4678]: I1013 12:44:48.996869 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:48Z","lastTransitionTime":"2025-10-13T12:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.048122 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/2.log" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.050946 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.051409 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.066932 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.087388 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.099471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.099589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.099607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.099639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.099657 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.110910 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.134125 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.145096 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.156779 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.168789 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.183456 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.196128 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.201831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.201871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.201883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.201901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.201916 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.211114 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.225109 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.240497 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.264726 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.278911 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.292479 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.304941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.305000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.305017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.305071 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.305090 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.306891 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.318475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.329987 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:49Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.406690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.406741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.406757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.406777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.406794 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.509844 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.509894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.509906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.509924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.509935 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.591710 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.591730 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.591726 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.591844 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:49 crc kubenswrapper[4678]: E1013 12:44:49.591961 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:49 crc kubenswrapper[4678]: E1013 12:44:49.592042 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:49 crc kubenswrapper[4678]: E1013 12:44:49.592229 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:49 crc kubenswrapper[4678]: E1013 12:44:49.592361 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.612299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.612350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.612367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.612389 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.612406 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.715537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.715594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.715610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.715633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.715685 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.818912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.818963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.818978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.818996 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.819010 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.921410 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.921448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.921457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.921471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:49 crc kubenswrapper[4678]: I1013 12:44:49.921481 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:49Z","lastTransitionTime":"2025-10-13T12:44:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.024156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.024194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.024204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.024217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.024225 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.056692 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/3.log" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.057430 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/2.log" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.060877 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" exitCode=1 Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.060913 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.060957 4678 scope.go:117] "RemoveContainer" containerID="feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.061591 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.061764 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.081820 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.092631 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.106721 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.116985 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.126199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.126234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.126245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.126260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.126272 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.128564 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.139121 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.150042 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.168634 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.196533 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.224645 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.227818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.227863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.227875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.227913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.227924 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.241379 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://feb6f02ce538e5ec9b9ce834b378afcb32ce64db7ce58e53e9a933c078277392\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:19Z\\\",\\\"message\\\":\\\"ault network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:19Z is after 2025-08-24T17:21:41Z]\\\\nI1013 12:44:19.691905 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:19.691922 6328 ovn.go:134] Ensuring zone local for Pod openshift-network-operator/iptables-alerter-4ln5h in node crc\\\\nI1013 12:44:19.691931 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1013 12:44:19.691824 6328 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nI1013 12:44:19.691940 6328 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-ale\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:49Z\\\",\\\"message\\\":\\\"44:49.548736 6693 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:49.548731 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548713 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548754 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548756 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548789 6693 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1013 12:44:49.548810 6693 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.251867 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.262223 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.270961 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.283172 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.293969 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.307556 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.315210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.315250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.315265 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.315282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.315293 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.318863 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.325040 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.328237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.328271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.328280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.328296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.328306 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.337831 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.340412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.340442 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.340452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.340466 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.340476 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.353903 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.393919 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.393948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.393956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.393971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.393980 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.403598 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.406644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.406672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.406682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.406696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.406705 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.416554 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:50Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:50 crc kubenswrapper[4678]: E1013 12:44:50.416665 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.417780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.417802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.417810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.417821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.417830 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.520475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.520531 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.520548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.520571 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.520588 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.623390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.623451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.623472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.623497 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.623515 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.725297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.725339 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.725356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.725376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.725393 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.827159 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.827189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.827197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.827211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.827219 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.929787 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.929823 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.929834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.929849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:50 crc kubenswrapper[4678]: I1013 12:44:50.929860 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:50Z","lastTransitionTime":"2025-10-13T12:44:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.032514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.032585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.032607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.032638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.032661 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.069384 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/3.log" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.073950 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:44:51 crc kubenswrapper[4678]: E1013 12:44:51.074187 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.092190 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.105764 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.117942 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.135487 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.135544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.135561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.135584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.135600 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.149539 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.168547 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.181421 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.201403 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.217853 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.235295 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.238195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.238405 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.238539 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.238691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.238832 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.248258 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.262248 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.279531 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.298468 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:49Z\\\",\\\"message\\\":\\\"44:49.548736 6693 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:49.548731 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548713 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548754 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548756 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548789 6693 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1013 12:44:49.548810 6693 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.307400 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.317787 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.327671 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.341512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.341562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.341573 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.341591 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.341602 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.343834 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.358883 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:51Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.444755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.444840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.444861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.444883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.444901 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.547866 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.547998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.548018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.548041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.548106 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.591970 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.592030 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:51 crc kubenswrapper[4678]: E1013 12:44:51.592231 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.592345 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:51 crc kubenswrapper[4678]: E1013 12:44:51.592477 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:51 crc kubenswrapper[4678]: E1013 12:44:51.592545 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.592603 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:51 crc kubenswrapper[4678]: E1013 12:44:51.592701 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.650426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.650486 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.650504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.650527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.650544 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.754498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.754579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.754601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.754629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.754647 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.857123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.857174 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.857185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.857205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.857216 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.959344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.959397 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.959414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.959438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:51 crc kubenswrapper[4678]: I1013 12:44:51.959456 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:51Z","lastTransitionTime":"2025-10-13T12:44:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.062314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.062379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.062397 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.062425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.062461 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.165910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.165959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.165970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.165987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.166000 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.269415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.269472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.269494 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.269523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.269546 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.373146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.373199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.373215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.373239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.373256 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.476221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.476285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.476304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.476327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.476375 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.579868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.579921 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.579938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.579963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.579980 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.613893 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.632609 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.645541 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.677307 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.682541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.682582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.682592 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.682610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.682622 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.698351 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.721714 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.741153 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.758695 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.778207 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.785969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.786013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.786030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.786083 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.786101 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.793587 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.810873 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.832029 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.863357 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:49Z\\\",\\\"message\\\":\\\"44:49.548736 6693 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:49.548731 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548713 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548754 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548756 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548789 6693 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1013 12:44:49.548810 6693 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.877500 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.890353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.890406 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.890424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.890447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.890467 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.892613 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.909551 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.924494 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.940451 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:44:52Z is after 2025-08-24T17:21:41Z" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.993194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.993250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.993267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.993291 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:52 crc kubenswrapper[4678]: I1013 12:44:52.993309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:52Z","lastTransitionTime":"2025-10-13T12:44:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.095393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.095432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.095443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.095460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.095472 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.198741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.198794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.198811 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.198834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.198850 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.302296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.302678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.302902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.303099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.303284 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.406512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.406577 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.406593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.406617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.406635 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.510207 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.510269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.510287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.510310 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.510327 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.591271 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:53 crc kubenswrapper[4678]: E1013 12:44:53.591403 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.591607 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:53 crc kubenswrapper[4678]: E1013 12:44:53.591668 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.591795 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:53 crc kubenswrapper[4678]: E1013 12:44:53.591850 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.591981 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:53 crc kubenswrapper[4678]: E1013 12:44:53.592044 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.613008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.613096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.613111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.613129 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.613141 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.716236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.716295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.716314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.716337 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.716354 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.819170 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.819221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.819243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.819271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.819292 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.922150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.922206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.922222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.922245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:53 crc kubenswrapper[4678]: I1013 12:44:53.922263 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:53Z","lastTransitionTime":"2025-10-13T12:44:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.025916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.025998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.026024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.026088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.026113 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.128818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.128886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.128904 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.128955 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.128974 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.232180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.232228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.232245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.232270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.232287 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.335114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.335177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.335195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.335226 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.335245 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.437943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.438003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.438023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.438078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.438097 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.541218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.541621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.541645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.541678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.541702 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.645226 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.645291 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.645314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.645342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.645363 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.748641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.748704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.748729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.748761 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.748782 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.851902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.851959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.851976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.852001 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.852019 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.955163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.955222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.955238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.955264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:54 crc kubenswrapper[4678]: I1013 12:44:54.955282 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:54Z","lastTransitionTime":"2025-10-13T12:44:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.058092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.058153 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.058169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.058194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.058211 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.161268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.161330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.161347 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.161373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.161391 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.263673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.263716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.263730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.263746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.263759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.367043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.367126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.367143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.367168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.367186 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.470337 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.470399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.470417 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.470441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.470458 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.572994 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.573108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.573135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.573169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.573192 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.591528 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.591611 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.591712 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.591739 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.591817 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.591876 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.591978 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.592075 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.676561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.676617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.676634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.676659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.676676 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.738037 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.738200 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.738171636 +0000 UTC m=+147.822709560 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.779304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.779359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.779376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.779399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.779419 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.839358 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.839419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.839475 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839656 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839681 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839690 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.839704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839740 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839763 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839853 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.839822294 +0000 UTC m=+147.924360238 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839887 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839701 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.839978 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.839955878 +0000 UTC m=+147.924493802 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.840007 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.839993779 +0000 UTC m=+147.924531693 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.840048 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: E1013 12:44:55.840134 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.840118682 +0000 UTC m=+147.924656606 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.882436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.882504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.882522 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.882546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.882564 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.984950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.985020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.985041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.985101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:55 crc kubenswrapper[4678]: I1013 12:44:55.985121 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:55Z","lastTransitionTime":"2025-10-13T12:44:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.088516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.088593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.088630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.088663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.088680 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.192189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.192245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.192263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.192287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.192307 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.295705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.295773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.295792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.295817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.295834 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.399280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.399330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.399349 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.399372 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.399390 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.502516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.502614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.502665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.502695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.502713 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.604639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.604704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.604723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.604747 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.604767 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.707557 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.707613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.707629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.707652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.707669 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.810854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.810919 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.810946 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.810977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.810998 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.914706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.914762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.914779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.914804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:56 crc kubenswrapper[4678]: I1013 12:44:56.914826 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:56Z","lastTransitionTime":"2025-10-13T12:44:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.017920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.018010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.018030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.018085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.018105 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.121122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.121155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.121164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.121177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.121187 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.224430 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.224492 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.224514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.224542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.224562 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.327505 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.327560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.327576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.327599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.327616 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.431395 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.431457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.431476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.431501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.431520 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.534820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.534949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.534967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.534993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.535036 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.592441 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.592527 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.592457 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.592441 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:57 crc kubenswrapper[4678]: E1013 12:44:57.592617 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:57 crc kubenswrapper[4678]: E1013 12:44:57.592722 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:57 crc kubenswrapper[4678]: E1013 12:44:57.592828 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:57 crc kubenswrapper[4678]: E1013 12:44:57.592913 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.641191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.641280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.641303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.641336 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.641358 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.743957 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.744013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.744029 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.744083 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.744101 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.847879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.847940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.847956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.847979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.847996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.951570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.951662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.951687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.951720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:57 crc kubenswrapper[4678]: I1013 12:44:57.951748 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:57Z","lastTransitionTime":"2025-10-13T12:44:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.054968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.055003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.055014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.055031 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.055043 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.157740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.157779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.157795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.157820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.157837 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.260639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.260728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.260754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.260790 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.260816 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.364558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.364606 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.364621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.364638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.364651 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.467154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.467215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.467232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.467256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.467274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.569969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.570038 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.570093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.570126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.570143 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.673484 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.673562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.673579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.673606 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.673626 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.776820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.776907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.776930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.776960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.776981 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.879971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.880040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.880090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.880122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.880146 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.984148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.984205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.984222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.984246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:58 crc kubenswrapper[4678]: I1013 12:44:58.984267 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:58Z","lastTransitionTime":"2025-10-13T12:44:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.087032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.087119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.087136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.087161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.087179 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.189126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.189190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.189207 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.189231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.189250 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.291795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.291864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.291882 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.291907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.291926 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.394243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.394324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.394345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.394371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.394390 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.498309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.498364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.498386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.498414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.498435 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.599912 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.600000 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.600154 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:44:59 crc kubenswrapper[4678]: E1013 12:44:59.600326 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.600435 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:44:59 crc kubenswrapper[4678]: E1013 12:44:59.600485 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:44:59 crc kubenswrapper[4678]: E1013 12:44:59.600615 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:44:59 crc kubenswrapper[4678]: E1013 12:44:59.600747 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.601230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.601277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.601296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.601319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.601339 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.703797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.703860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.703878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.703903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.703921 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.806696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.806797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.806819 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.806844 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.806866 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.910000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.910086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.910107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.910131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:44:59 crc kubenswrapper[4678]: I1013 12:44:59.910148 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:44:59Z","lastTransitionTime":"2025-10-13T12:44:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.012742 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.012809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.012828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.012855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.012878 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.116505 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.116567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.116584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.116612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.116635 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.219323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.219355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.219364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.219377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.219385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.321616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.321650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.321660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.321673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.321682 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.423924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.423991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.424018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.424046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.424094 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.527322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.527378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.527396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.527420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.527439 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.630962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.631118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.631144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.631174 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.631196 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.647940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.648024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.648040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.648082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.648102 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.667847 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.673009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.673147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.673166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.673188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.673205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.689621 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.693850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.693895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.693905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.693921 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.693934 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.716805 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.721709 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.721740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.721750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.721766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.721777 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.737423 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.741076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.741128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.741141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.741155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.741471 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.756467 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:00Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:00 crc kubenswrapper[4678]: E1013 12:45:00.756671 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.757858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.757884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.757893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.757908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.757922 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.860555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.860627 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.860650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.860679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.860701 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.963402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.963460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.963476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.963501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:00 crc kubenswrapper[4678]: I1013 12:45:00.963525 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:00Z","lastTransitionTime":"2025-10-13T12:45:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.066345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.066452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.066478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.066507 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.066528 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.168946 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.169003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.169019 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.169040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.169201 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.271985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.272086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.272106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.272131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.272151 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.375280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.375390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.375414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.375441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.375456 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.478405 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.478450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.478461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.478480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.478491 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.581545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.581607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.581618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.581640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.581660 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.591843 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:01 crc kubenswrapper[4678]: E1013 12:45:01.591980 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.592190 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:01 crc kubenswrapper[4678]: E1013 12:45:01.592267 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.592408 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:01 crc kubenswrapper[4678]: E1013 12:45:01.592476 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.592613 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:01 crc kubenswrapper[4678]: E1013 12:45:01.592687 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.683869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.683931 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.683952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.683982 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.684005 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.786652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.786699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.786718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.786740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.786759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.890237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.890304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.890323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.890352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.890370 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.993880 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.993951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.993976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.994004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:01 crc kubenswrapper[4678]: I1013 12:45:01.994027 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:01Z","lastTransitionTime":"2025-10-13T12:45:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.096758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.096815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.096833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.096856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.096874 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.200096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.200157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.200173 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.200196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.200212 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.303789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.303851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.303869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.303896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.303915 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.407301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.407363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.407380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.407405 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.407422 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.510198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.510256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.510274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.510301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.510320 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.593637 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:45:02 crc kubenswrapper[4678]: E1013 12:45:02.594001 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.615358 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.615930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.615968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.615988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.616011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.616028 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.635779 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.650264 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.668219 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.685766 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.717229 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.718462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.718520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.718541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.718568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.718589 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.738272 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.756806 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.775239 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.789925 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.803163 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.817609 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.821739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.821759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.821767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.821783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.821793 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.830586 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.845991 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.866959 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.897123 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:49Z\\\",\\\"message\\\":\\\"44:49.548736 6693 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:49.548731 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548713 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548754 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548756 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548789 6693 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1013 12:44:49.548810 6693 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.919288 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.924277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.924345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.924371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.924403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.924428 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:02Z","lastTransitionTime":"2025-10-13T12:45:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:02 crc kubenswrapper[4678]: I1013 12:45:02.939476 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:02Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.027565 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.027615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.027630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.027651 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.027664 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.130327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.130396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.130420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.130450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.130473 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.233006 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.233040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.233067 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.233082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.233093 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.335965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.336041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.336076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.336099 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.336112 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.438705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.438770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.438791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.438818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.438844 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.542042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.542112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.542126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.542145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.542159 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.592116 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.592186 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.592186 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.592129 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:03 crc kubenswrapper[4678]: E1013 12:45:03.592303 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:03 crc kubenswrapper[4678]: E1013 12:45:03.592425 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:03 crc kubenswrapper[4678]: E1013 12:45:03.592543 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:03 crc kubenswrapper[4678]: E1013 12:45:03.592593 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.645365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.645412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.645425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.645443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.645457 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.749117 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.749177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.749202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.749233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.749257 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.852653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.852716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.852733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.852759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.852780 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.955826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.955902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.955921 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.955947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:03 crc kubenswrapper[4678]: I1013 12:45:03.955967 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:03Z","lastTransitionTime":"2025-10-13T12:45:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.060171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.060290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.060311 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.060784 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.060843 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.163573 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.163608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.163618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.163631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.163641 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.265740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.265781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.265792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.265838 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.265853 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.369295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.369350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.369364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.369384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.369400 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.471516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.471556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.471568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.471586 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.471599 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.574228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.574268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.574281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.574297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.574308 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.677211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.677274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.677295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.677320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.677339 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.779790 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.779841 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.779856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.779881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.779897 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.882514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.882554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.882563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.882578 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.882587 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.985834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.985873 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.985881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.985895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:04 crc kubenswrapper[4678]: I1013 12:45:04.985905 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:04Z","lastTransitionTime":"2025-10-13T12:45:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.088864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.088934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.088960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.088993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.089015 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.192153 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.192213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.192229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.192251 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.192269 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.295620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.295698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.295716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.295742 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.295760 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.399114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.399196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.399215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.399238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.399257 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.502728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.502824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.502850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.502888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.502912 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.591326 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.591357 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.591677 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:05 crc kubenswrapper[4678]: E1013 12:45:05.591713 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.591798 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:05 crc kubenswrapper[4678]: E1013 12:45:05.591581 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:05 crc kubenswrapper[4678]: E1013 12:45:05.592414 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:05 crc kubenswrapper[4678]: E1013 12:45:05.592552 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.606971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.607037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.607086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.607128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.607173 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.607515 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.710588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.710641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.710661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.710683 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.710699 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.813028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.813125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.813145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.813165 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.813182 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.915746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.915819 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.915843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.915871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:05 crc kubenswrapper[4678]: I1013 12:45:05.915887 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:05Z","lastTransitionTime":"2025-10-13T12:45:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.018542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.018616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.018637 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.018666 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.018687 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.121773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.121842 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.121885 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.121915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.121938 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.224445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.224497 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.224514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.224538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.224561 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.327585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.327643 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.327661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.327685 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.327704 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.430036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.430132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.430155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.430183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.430205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.533154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.533204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.533222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.533243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.533258 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.635653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.635722 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.635739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.635764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.635785 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.738845 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.738908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.738932 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.738963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.738986 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.841845 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.841906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.841923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.841948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.841966 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.944905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.944956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.944973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.944996 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:06 crc kubenswrapper[4678]: I1013 12:45:06.945012 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:06Z","lastTransitionTime":"2025-10-13T12:45:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.048552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.048610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.048628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.048652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.048671 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.150607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.150690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.150715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.150748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.150770 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.254322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.254427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.254453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.254484 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.254505 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.357503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.357580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.357601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.357633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.357656 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.460941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.461030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.461083 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.461115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.461136 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.564489 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.564546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.564564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.564590 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.564607 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.592231 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.592306 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.592365 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:07 crc kubenswrapper[4678]: E1013 12:45:07.592445 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.592507 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:07 crc kubenswrapper[4678]: E1013 12:45:07.592576 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:07 crc kubenswrapper[4678]: E1013 12:45:07.592774 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:07 crc kubenswrapper[4678]: E1013 12:45:07.592923 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.667687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.667736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.667756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.667781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.667798 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.771068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.771102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.771116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.771131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.771142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.874245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.874304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.874320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.874344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.874361 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.978584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.978649 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.978665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.978691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:07 crc kubenswrapper[4678]: I1013 12:45:07.978710 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:07Z","lastTransitionTime":"2025-10-13T12:45:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.081144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.081212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.081229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.081256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.081273 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.184034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.184127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.184145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.184171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.184189 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.287985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.288025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.288038 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.288075 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.288088 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.390861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.390920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.390938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.390969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.390993 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.494630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.494715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.494734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.494762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.494781 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.597453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.597542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.597566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.597595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.597612 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.700745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.700808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.700825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.700848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.700865 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.803509 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.803583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.803600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.803624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.803641 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.906305 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.906371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.906383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.906401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:08 crc kubenswrapper[4678]: I1013 12:45:08.906412 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:08Z","lastTransitionTime":"2025-10-13T12:45:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.010688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.010749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.010766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.010791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.010807 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.112663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.112713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.112726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.112744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.112758 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.214850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.214912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.214929 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.214952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.214969 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.317831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.317892 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.317910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.317933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.317950 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.420717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.420769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.420781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.420802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.420815 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.523304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.523341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.523352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.523368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.523380 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.592120 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.592208 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:09 crc kubenswrapper[4678]: E1013 12:45:09.592232 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.592297 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.592353 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:09 crc kubenswrapper[4678]: E1013 12:45:09.592512 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:09 crc kubenswrapper[4678]: E1013 12:45:09.592642 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:09 crc kubenswrapper[4678]: E1013 12:45:09.592758 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.626361 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.626429 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.626456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.626486 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.626508 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.729213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.729281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.729297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.729320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.729337 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.832863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.832922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.832941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.832964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.832983 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.935243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.935320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.935345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.935374 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:09 crc kubenswrapper[4678]: I1013 12:45:09.935397 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:09Z","lastTransitionTime":"2025-10-13T12:45:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.038202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.038298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.038317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.038343 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.038362 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.140815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.140922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.140950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.140981 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.141003 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.244523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.244605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.244632 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.244665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.244690 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.348762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.348818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.348830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.348851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.348864 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.451763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.451810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.451820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.451836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.451848 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.555177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.555247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.555268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.555293 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.555312 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.658470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.658537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.658556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.658580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.658600 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.761712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.761774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.761791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.761815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.761833 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.804438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.804487 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.804508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.804530 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.804545 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.824528 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.830030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.830130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.830154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.830183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.830204 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.849674 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.855309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.855357 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.855374 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.855396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.855413 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.881130 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.886400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.886508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.887039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.887147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.887167 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.908895 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.913835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.913903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.913925 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.913955 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.913979 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.935030 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-13T12:45:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4d85342c-260c-43e8-af13-3624a88b6f98\\\",\\\"systemUUID\\\":\\\"bc86650e-4c93-41c1-ba1f-a59fdc4063ca\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:10Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:10 crc kubenswrapper[4678]: E1013 12:45:10.935326 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.937766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.937818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.937836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.937860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:10 crc kubenswrapper[4678]: I1013 12:45:10.937878 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:10Z","lastTransitionTime":"2025-10-13T12:45:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.040725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.040801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.040825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.040857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.040880 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.144504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.144583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.144606 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.144631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.144649 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.248348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.248426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.248450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.248477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.248502 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.351696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.351771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.351795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.351830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.351855 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.455152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.455216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.455233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.455256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.455273 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.558405 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.558503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.558516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.558534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.558547 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.591801 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.591861 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.591929 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.592004 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.592099 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.592163 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.592271 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.592450 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.661739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.661816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.661840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.661870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.661893 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.765457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.765518 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.765539 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.765568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.765586 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.868149 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.868216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.868233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.868260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.868279 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.940936 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.941201 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:45:11 crc kubenswrapper[4678]: E1013 12:45:11.941292 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs podName:4ccdd1a7-2f11-43fd-9369-5ca8380ef70a nodeName:}" failed. No retries permitted until 2025-10-13 12:46:15.94126928 +0000 UTC m=+164.025807204 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs") pod "network-metrics-daemon-hbs76" (UID: "4ccdd1a7-2f11-43fd-9369-5ca8380ef70a") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.970664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.970741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.970759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.970785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:11 crc kubenswrapper[4678]: I1013 12:45:11.970802 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:11Z","lastTransitionTime":"2025-10-13T12:45:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.073919 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.073967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.073987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.074012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.074032 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.176821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.176895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.176912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.176937 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.176956 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.279461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.279503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.279516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.279533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.279545 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.381906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.381970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.381992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.382020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.382041 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.484687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.484749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.484771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.484801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.484825 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.587855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.587922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.587942 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.587968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.587986 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.610526 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"55fd7175-35d2-439c-af7c-4fff7b766061\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://de74300447ee49ec05aba18d895058af13c1ca2fc4ac500f5249d7ccd163c8a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73781957fc5d35902182dbd31ec90b4254d867825c76296ddae10d6f96ddb333\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73781957fc5d35902182dbd31ec90b4254d867825c76296ddae10d6f96ddb333\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.632664 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c020e3ea94dcce1730f51f1a4f9f044e2f67acfe65949f3aeec64c38b67b69e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.654525 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.669670 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-925jf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6539fec7-6f19-4758-8da3-13a6f978d5a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3df8bb70f8b2b26e6849f778d2486ee2c3df0d0aeb0cf1b9fe8424c4198ccf25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j6sdx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-925jf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.691183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.691254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.691277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.691312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.691488 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.705366 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dd3074d2-f339-4d68-b600-030d90003efc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab864f0b85b5aeada37ab3ca61b25b738fa8b92afc9f40d53590732d164114da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e39e0b5573ba00b3d7e197df6eb4fe26ff2c1db03765b5612bd596aee88d320f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd69db8ed06d1622b9ec55e4e64bcbd93d8e20c67979af39ac07430a77dfaac2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f7ebf71b0cfcd472ef70e60d0314cc0eb8544b06ee349db0b92d350af81c631\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bdd9cdaf2c830e7dafa07252c29c0ef0b46f4620ed4a11ce262ef2e13b4155ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e13341d02841ad8ad7d63786fb84d51129bb023c31f6168d21e38cea2a0884d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1b628989da2e5026a5e76c9b8e4629188fc6c049d13d85c6ad94ccf8bb60c940\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3560b636e5ccc314a38411f9c47657e4a9ae99d012871bcea74a80693946735\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.726681 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6c6c51f1-60df-4ed3-8793-0fcf7efa4057\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd41712a94fa3f7186af2bddf5bbaef3a42ce8d1e4f197da9c8ac173e9e33063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e339029c6b10dc3856eadfad04ca07f4d6329077b1a300df6c46a69f43c5e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab98cd9e8d7f9326e82b311f301ec2f2aab5a8242e6fb37494a86352103dc11a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ae2dee45bf68355047a418e8d9e06c4591da1f2dcc62990a9d537af5fa7e817\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.747548 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.767813 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21e458ad3f9dda90d111c1080d1783dfd877eefcbbe60562ed459ab3bee4fa06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71693f2f62a25da8fb361b226b5c3b724e42470968121f1f2d0a8434a168f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.787748 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ac3ce51f360ea211d07e32c8d06b5e946b51a0f87e932ee4fa841fe21d52917\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.794130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.794172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.794190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.794212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.794229 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.808853 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-nlfg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21d5191d-621a-4416-b635-e3459b3cfe8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:40Z\\\",\\\"message\\\":\\\"2025-10-13T12:43:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9\\\\n2025-10-13T12:43:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8759245c-9540-443a-9719-d458244ca5d9 to /host/opt/cni/bin/\\\\n2025-10-13T12:43:55Z [verbose] multus-daemon started\\\\n2025-10-13T12:43:55Z [verbose] Readiness Indicator file check\\\\n2025-10-13T12:44:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w95kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-nlfg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.825883 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbs76" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-47mr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbs76\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.842532 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b63499c-aef5-4851-bf32-920aee3bc2a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81f20c4055d547b104c3b5624fca9b34dd7bb22c8bb713e41fc8384e39d97cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f22b8a7cb13dfc583cee83ac1f355b300b47936255462ae88742ee89da485\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d9b42979ab7b99b4df843d48fc64b806eebd5807bb1b1a883dc85097fb86119\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d08810e01945f20b015bdee7e6008ab6a5c12429afaccfee142da98eab1cf100\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.861303 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c6e2144-3790-4a4a-b47c-5b3459dacdf9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://932db0d4244ba0162ecefe049524f879175c77294932278d5e9b537663ff7637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xxp2g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bxtwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.883555 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2dee984d-00a7-4d89-a39f-1cfaac6bb2f6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdc2ed0283263f389e5c23ed6a81f89fa78c1c6c09b8b0109e5817794fef9858\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bac9d2b46ab88570cd11d9e9f79ac979ddd016eccfccb528407e107ce6f98dd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dadf2e42e850d7b4b7bd4a7074b644f39789955989fd21252c74502d2a794e6b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ee1e71687cff0faae7d543764824dc14551644162993eb71403a3cea3416a37\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62771f78938776979adb715ae0700a4216e30b6cf0a5a0aa9a1719434aa7bcb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d250a971a9a7220c88b37d92c01ed22832d1ba8ea2af437b9dd40774faa0fec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b2d84ae70589ceb8bc6b41a1cb25b07bd99b2c5467a80b2fc30406739c89ef6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzbsl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mbz2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.897938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.898081 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.898103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.898127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.898148 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:12Z","lastTransitionTime":"2025-10-13T12:45:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.913246 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"43e78486-f760-44e0-a85e-f3868869ae5b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-13T12:44:49Z\\\",\\\"message\\\":\\\"44:49.548736 6693 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1013 12:44:49.548731 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548713 6693 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548754 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-hbs76\\\\nI1013 12:44:49.548756 6693 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1013 12:44:49.548789 6693 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF1013 12:44:49.548810 6693 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal e\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-13T12:44:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q4tf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hsx89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.928277 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-57x29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f51fa35-1679-4abf-8d05-1070830d0f42\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52a0e8c1f0c31b6a448198d5f495a61f2852e68b6d5a8e797dd894609517fa38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7x5p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-57x29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.944020 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1c320f8-4efb-44fd-8ad0-fa27faea2811\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:44:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6936a15a5306ff5f792ccac681521c19168b3ed675d155ea82da7a1c50c9a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3777800a9e5cf0035afeb71b9215602eaceeec7e99f159d66ad6123165d6bb91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:44:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-475dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:44:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-n5l5k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.963207 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5301e1ad-4725-4690-8c43-b0be053693f2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b53c07a713ef24dea93712d388f8701318e262d2dfa7b139a9314156ade9f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3cbb9640bfb18b4fa47734a8216ccaae5ef6cc35e259d17af91f97d8bd363d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05e6091fd0e58b301672d9bf32616a272f8012f402b93ab180dccf208755a696\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1a035000ee4ad40fd8e8b571e951386288d9c5bad99657d60a027503684e4c4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a8f5ca110264a899260e32ee69fc63d5f62e2cd9d84b175b26457c2bbb79c5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-13T12:43:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e614a70e047a6fd58f5e1cbb481ec6d1f9ca92dc77f104a45ed2048a1cfd1557\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-13T12:43:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-13T12:43:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-13T12:43:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:12 crc kubenswrapper[4678]: I1013 12:45:12.979954 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-13T12:43:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-13T12:45:12Z is after 2025-08-24T17:21:41Z" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.001212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.001260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.001278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.001299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.001316 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.104288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.104336 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.104395 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.104427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.104443 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.206563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.206729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.206758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.206788 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.206820 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.309474 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.309536 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.309554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.309578 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.309597 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.413085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.413146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.413179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.413199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.413211 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.515794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.515867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.515886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.515916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.515937 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.591846 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.591894 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.591981 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:13 crc kubenswrapper[4678]: E1013 12:45:13.592124 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.592040 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:13 crc kubenswrapper[4678]: E1013 12:45:13.592024 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:13 crc kubenswrapper[4678]: E1013 12:45:13.592231 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:13 crc kubenswrapper[4678]: E1013 12:45:13.592394 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.623426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.623504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.623524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.623551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.623570 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.727495 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.727556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.727573 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.727598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.727615 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.830853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.830970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.830988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.831010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.831026 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.934740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.934792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.934809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.934834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:13 crc kubenswrapper[4678]: I1013 12:45:13.934853 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:13Z","lastTransitionTime":"2025-10-13T12:45:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.038134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.038201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.038220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.038651 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.038706 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.142622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.142689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.142713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.142742 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.143010 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.246941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.247015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.247037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.247110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.247135 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.349817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.349887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.349909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.349943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.349962 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.452667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.452733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.452749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.452773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.452790 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.555512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.555571 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.555587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.555612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.555632 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.659194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.659287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.659304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.659327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.659346 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.762897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.762977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.762995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.763017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.763033 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.865659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.865712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.865729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.865753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.865769 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.969313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.969375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.969398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.969426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:14 crc kubenswrapper[4678]: I1013 12:45:14.969447 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:14Z","lastTransitionTime":"2025-10-13T12:45:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.076585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.076675 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.076697 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.076725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.076748 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.179418 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.179478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.179496 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.179521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.179540 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.282355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.282420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.282437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.282462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.282478 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.385354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.385409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.385427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.385449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.385467 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.488221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.488321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.488366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.488391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.488408 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591177 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591281 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591319 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:15 crc kubenswrapper[4678]: E1013 12:45:15.591422 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591516 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.591561 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:15 crc kubenswrapper[4678]: E1013 12:45:15.591674 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:15 crc kubenswrapper[4678]: E1013 12:45:15.591856 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:15 crc kubenswrapper[4678]: E1013 12:45:15.591987 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.593344 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:45:15 crc kubenswrapper[4678]: E1013 12:45:15.593654 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hsx89_openshift-ovn-kubernetes(43e78486-f760-44e0-a85e-f3868869ae5b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.694444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.694507 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.694524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.694549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.694566 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.797579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.797647 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.797662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.797687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.797709 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.900551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.900611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.900633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.900700 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:15 crc kubenswrapper[4678]: I1013 12:45:15.900723 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:15Z","lastTransitionTime":"2025-10-13T12:45:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.003410 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.003511 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.003534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.003568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.003591 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.106987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.107108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.107128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.107152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.107168 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.216345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.216424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.216445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.216470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.216496 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.319436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.319498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.319512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.319525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.319534 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.422113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.422166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.422177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.422191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.422203 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.525228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.525262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.525276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.525290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.525300 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.627595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.627634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.627642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.627657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.627666 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.730851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.730899 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.730914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.730936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.730952 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.833807 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.833865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.833883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.833905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.833919 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.936960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.937021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.937104 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.937132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:16 crc kubenswrapper[4678]: I1013 12:45:16.937151 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:16Z","lastTransitionTime":"2025-10-13T12:45:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.040169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.040451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.040607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.040747 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.040884 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.143513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.143588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.143613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.143644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.143662 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.246328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.246528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.246674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.246842 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.246992 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.349843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.350765 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.350908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.351032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.351208 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.454367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.454430 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.454449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.454473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.454491 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.557670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.557744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.557766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.557793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.557816 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.592274 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.592312 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.592376 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:17 crc kubenswrapper[4678]: E1013 12:45:17.592494 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:17 crc kubenswrapper[4678]: E1013 12:45:17.592601 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:17 crc kubenswrapper[4678]: E1013 12:45:17.592717 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.592820 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:17 crc kubenswrapper[4678]: E1013 12:45:17.613562 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.660448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.660491 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.660501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.660520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.660532 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.763444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.763501 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.763519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.763547 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.763565 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.867318 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.867385 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.867409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.867441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.867467 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.970398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.970471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.970494 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.970530 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:17 crc kubenswrapper[4678]: I1013 12:45:17.970554 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:17Z","lastTransitionTime":"2025-10-13T12:45:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.073629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.073686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.073703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.073727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.073747 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.176824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.176897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.176915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.176942 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.176959 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.280315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.280381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.280398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.280422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.280439 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.383433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.383509 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.383527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.383556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.383576 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.486693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.486749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.486767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.486789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.486807 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.589770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.589842 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.589859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.590300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.590359 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.693749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.693814 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.693831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.693858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.693875 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.796945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.797009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.797028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.797078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.797097 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.900141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.900218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.900241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.900272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:18 crc kubenswrapper[4678]: I1013 12:45:18.900297 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:18Z","lastTransitionTime":"2025-10-13T12:45:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.003524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.003596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.003613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.003635 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.003654 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.106765 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.106843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.106873 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.106902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.106923 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.215017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.215111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.215132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.215160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.215179 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.318272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.318329 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.318347 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.318371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.318389 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.421785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.421841 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.421863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.421886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.421904 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.524934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.524997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.525014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.525039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.525083 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.591391 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:19 crc kubenswrapper[4678]: E1013 12:45:19.591570 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.591611 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.591626 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.591939 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:19 crc kubenswrapper[4678]: E1013 12:45:19.592102 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:19 crc kubenswrapper[4678]: E1013 12:45:19.592237 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:19 crc kubenswrapper[4678]: E1013 12:45:19.592515 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.628342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.628407 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.628424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.628452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.628472 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.732178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.732238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.732255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.732278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.732294 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.835956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.836085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.836117 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.836150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.836175 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.939296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.939365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.939389 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.939419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:19 crc kubenswrapper[4678]: I1013 12:45:19.939439 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:19Z","lastTransitionTime":"2025-10-13T12:45:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.041935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.042047 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.042088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.042116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.042137 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.145368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.145419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.145435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.145459 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.145475 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.248628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.248705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.248723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.248752 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.248771 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.351612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.351681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.351699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.351726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.351744 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.454784 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.454828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.454838 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.454852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.454861 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.558324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.558391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.558415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.558483 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.558507 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.660573 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.660642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.660665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.660692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.660714 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.763792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.763884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.763901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.763921 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.763938 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.866617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.866687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.866707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.866732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.866750 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.969481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.969533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.969549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.969572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:20 crc kubenswrapper[4678]: I1013 12:45:20.969592 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:20Z","lastTransitionTime":"2025-10-13T12:45:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.072867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.072959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.072986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.073014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.073033 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:21Z","lastTransitionTime":"2025-10-13T12:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.157739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.157812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.157830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.157858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.157876 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:21Z","lastTransitionTime":"2025-10-13T12:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.193123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.193231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.193263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.193339 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.193363 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-13T12:45:21Z","lastTransitionTime":"2025-10-13T12:45:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.229793 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw"] Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.230807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.235443 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.235719 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.235451 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.235506 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.265192 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.265150992 podStartE2EDuration="56.265150992s" podCreationTimestamp="2025-10-13 12:44:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.2576358 +0000 UTC m=+109.342173724" watchObservedRunningTime="2025-10-13 12:45:21.265150992 +0000 UTC m=+109.349688916" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.279523 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podStartSLOduration=88.279493709 podStartE2EDuration="1m28.279493709s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.279274613 +0000 UTC m=+109.363812557" watchObservedRunningTime="2025-10-13 12:45:21.279493709 +0000 UTC m=+109.364031633" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.347218 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9mbz2" podStartSLOduration=88.347184189 podStartE2EDuration="1m28.347184189s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.310262255 +0000 UTC m=+109.394800179" watchObservedRunningTime="2025-10-13 12:45:21.347184189 +0000 UTC m=+109.431722113" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.348550 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.348841 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.348894 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-service-ca\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.348930 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.349009 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.362148 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-57x29" podStartSLOduration=89.362125161 podStartE2EDuration="1m29.362125161s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.360391986 +0000 UTC m=+109.444929910" watchObservedRunningTime="2025-10-13 12:45:21.362125161 +0000 UTC m=+109.446663055" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.374699 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-n5l5k" podStartSLOduration=88.374674651 podStartE2EDuration="1m28.374674651s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.373641355 +0000 UTC m=+109.458179279" watchObservedRunningTime="2025-10-13 12:45:21.374674651 +0000 UTC m=+109.459212565" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.427378 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.427358638 podStartE2EDuration="1m30.427358638s" podCreationTimestamp="2025-10-13 12:43:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.427098231 +0000 UTC m=+109.511636145" watchObservedRunningTime="2025-10-13 12:45:21.427358638 +0000 UTC m=+109.511896532" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450461 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450550 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-service-ca\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450603 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450638 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450756 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.450926 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.452095 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-service-ca\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.458733 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.461524 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.461505321 podStartE2EDuration="16.461505321s" podCreationTimestamp="2025-10-13 12:45:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.45833477 +0000 UTC m=+109.542872694" watchObservedRunningTime="2025-10-13 12:45:21.461505321 +0000 UTC m=+109.546043235" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.470094 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbc1cab0-5545-4c02-bbe2-a09a06d26f66-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-drnjw\" (UID: \"cbc1cab0-5545-4c02-bbe2-a09a06d26f66\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.517236 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-925jf" podStartSLOduration=89.517211504 podStartE2EDuration="1m29.517211504s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.516356432 +0000 UTC m=+109.600894346" watchObservedRunningTime="2025-10-13 12:45:21.517211504 +0000 UTC m=+109.601749398" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.546004 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=85.545980619 podStartE2EDuration="1m25.545980619s" podCreationTimestamp="2025-10-13 12:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.544953753 +0000 UTC m=+109.629491657" watchObservedRunningTime="2025-10-13 12:45:21.545980619 +0000 UTC m=+109.630518543" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.556011 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.591877 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.591877 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.591926 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.592030 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:21 crc kubenswrapper[4678]: E1013 12:45:21.592221 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:21 crc kubenswrapper[4678]: E1013 12:45:21.592330 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:21 crc kubenswrapper[4678]: E1013 12:45:21.592479 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:21 crc kubenswrapper[4678]: E1013 12:45:21.592607 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.593721 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=83.593704669 podStartE2EDuration="1m23.593704669s" podCreationTimestamp="2025-10-13 12:43:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.567354886 +0000 UTC m=+109.651892780" watchObservedRunningTime="2025-10-13 12:45:21.593704669 +0000 UTC m=+109.678242583" Oct 13 12:45:21 crc kubenswrapper[4678]: I1013 12:45:21.639499 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nlfg7" podStartSLOduration=88.639478549 podStartE2EDuration="1m28.639478549s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:21.63874073 +0000 UTC m=+109.723278664" watchObservedRunningTime="2025-10-13 12:45:21.639478549 +0000 UTC m=+109.724016443" Oct 13 12:45:22 crc kubenswrapper[4678]: I1013 12:45:22.185994 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" event={"ID":"cbc1cab0-5545-4c02-bbe2-a09a06d26f66","Type":"ContainerStarted","Data":"42c348564f634c6d1ee1c1908f73e0d63113ffc2a22ea791b45fd371dc072714"} Oct 13 12:45:22 crc kubenswrapper[4678]: I1013 12:45:22.186285 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" event={"ID":"cbc1cab0-5545-4c02-bbe2-a09a06d26f66","Type":"ContainerStarted","Data":"7075f5b4065926d0bd67259ba97d1203ab10a80ab8607d505c79742db3757606"} Oct 13 12:45:23 crc kubenswrapper[4678]: I1013 12:45:23.592203 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:23 crc kubenswrapper[4678]: I1013 12:45:23.592239 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:23 crc kubenswrapper[4678]: I1013 12:45:23.592344 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:23 crc kubenswrapper[4678]: E1013 12:45:23.592503 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:23 crc kubenswrapper[4678]: I1013 12:45:23.592559 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:23 crc kubenswrapper[4678]: E1013 12:45:23.592700 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:23 crc kubenswrapper[4678]: E1013 12:45:23.592810 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:23 crc kubenswrapper[4678]: E1013 12:45:23.592940 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:25 crc kubenswrapper[4678]: I1013 12:45:25.591856 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:25 crc kubenswrapper[4678]: I1013 12:45:25.591852 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:25 crc kubenswrapper[4678]: I1013 12:45:25.591862 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:25 crc kubenswrapper[4678]: I1013 12:45:25.591980 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:25 crc kubenswrapper[4678]: E1013 12:45:25.592143 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:25 crc kubenswrapper[4678]: E1013 12:45:25.592294 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:25 crc kubenswrapper[4678]: E1013 12:45:25.592448 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:25 crc kubenswrapper[4678]: E1013 12:45:25.592579 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.205959 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/1.log" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.206689 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/0.log" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.206761 4678 generic.go:334] "Generic (PLEG): container finished" podID="21d5191d-621a-4416-b635-e3459b3cfe8d" containerID="1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c" exitCode=1 Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.206793 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerDied","Data":"1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c"} Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.206831 4678 scope.go:117] "RemoveContainer" containerID="43715e3286e5c7fb81937ac50a4bfefddecfa49ddb8134915143ac03fa73e376" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.207475 4678 scope.go:117] "RemoveContainer" containerID="1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c" Oct 13 12:45:27 crc kubenswrapper[4678]: E1013 12:45:27.207766 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-nlfg7_openshift-multus(21d5191d-621a-4416-b635-e3459b3cfe8d)\"" pod="openshift-multus/multus-nlfg7" podUID="21d5191d-621a-4416-b635-e3459b3cfe8d" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.236677 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-drnjw" podStartSLOduration=94.236651308 podStartE2EDuration="1m34.236651308s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:22.202635112 +0000 UTC m=+110.287173016" watchObservedRunningTime="2025-10-13 12:45:27.236651308 +0000 UTC m=+115.321189222" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.591800 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.591848 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.591845 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:27 crc kubenswrapper[4678]: I1013 12:45:27.591908 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:27 crc kubenswrapper[4678]: E1013 12:45:27.591987 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:27 crc kubenswrapper[4678]: E1013 12:45:27.592156 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:27 crc kubenswrapper[4678]: E1013 12:45:27.592340 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:27 crc kubenswrapper[4678]: E1013 12:45:27.592443 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:28 crc kubenswrapper[4678]: I1013 12:45:28.211681 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/1.log" Oct 13 12:45:29 crc kubenswrapper[4678]: I1013 12:45:29.591311 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:29 crc kubenswrapper[4678]: E1013 12:45:29.591430 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:29 crc kubenswrapper[4678]: I1013 12:45:29.591469 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:29 crc kubenswrapper[4678]: I1013 12:45:29.591530 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:29 crc kubenswrapper[4678]: I1013 12:45:29.591567 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:29 crc kubenswrapper[4678]: E1013 12:45:29.591795 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:29 crc kubenswrapper[4678]: E1013 12:45:29.591957 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:29 crc kubenswrapper[4678]: E1013 12:45:29.592000 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:29 crc kubenswrapper[4678]: I1013 12:45:29.592550 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.221650 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/3.log" Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.226565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerStarted","Data":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.227167 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.278260 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podStartSLOduration=97.278238182 podStartE2EDuration="1m37.278238182s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:30.277636247 +0000 UTC m=+118.362174161" watchObservedRunningTime="2025-10-13 12:45:30.278238182 +0000 UTC m=+118.362776106" Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.583247 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hbs76"] Oct 13 12:45:30 crc kubenswrapper[4678]: I1013 12:45:30.583404 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:30 crc kubenswrapper[4678]: E1013 12:45:30.583534 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:31 crc kubenswrapper[4678]: I1013 12:45:31.592386 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:31 crc kubenswrapper[4678]: I1013 12:45:31.592565 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:31 crc kubenswrapper[4678]: I1013 12:45:31.592628 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:31 crc kubenswrapper[4678]: E1013 12:45:31.592970 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:31 crc kubenswrapper[4678]: E1013 12:45:31.593109 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:31 crc kubenswrapper[4678]: E1013 12:45:31.593234 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:32 crc kubenswrapper[4678]: E1013 12:45:32.527323 4678 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 13 12:45:32 crc kubenswrapper[4678]: I1013 12:45:32.592103 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:32 crc kubenswrapper[4678]: E1013 12:45:32.594361 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:32 crc kubenswrapper[4678]: E1013 12:45:32.705041 4678 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 13 12:45:33 crc kubenswrapper[4678]: I1013 12:45:33.591476 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:33 crc kubenswrapper[4678]: I1013 12:45:33.591526 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:33 crc kubenswrapper[4678]: E1013 12:45:33.591665 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:33 crc kubenswrapper[4678]: I1013 12:45:33.591540 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:33 crc kubenswrapper[4678]: E1013 12:45:33.591836 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:33 crc kubenswrapper[4678]: E1013 12:45:33.591941 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:34 crc kubenswrapper[4678]: I1013 12:45:34.591772 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:34 crc kubenswrapper[4678]: E1013 12:45:34.592387 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:35 crc kubenswrapper[4678]: I1013 12:45:35.592173 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:35 crc kubenswrapper[4678]: I1013 12:45:35.592235 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:35 crc kubenswrapper[4678]: I1013 12:45:35.592231 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:35 crc kubenswrapper[4678]: E1013 12:45:35.592438 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:35 crc kubenswrapper[4678]: E1013 12:45:35.592582 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:35 crc kubenswrapper[4678]: E1013 12:45:35.592706 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:36 crc kubenswrapper[4678]: I1013 12:45:36.592421 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:36 crc kubenswrapper[4678]: E1013 12:45:36.592638 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:37 crc kubenswrapper[4678]: I1013 12:45:37.592160 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:37 crc kubenswrapper[4678]: I1013 12:45:37.592205 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:37 crc kubenswrapper[4678]: I1013 12:45:37.592181 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:37 crc kubenswrapper[4678]: E1013 12:45:37.592348 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:37 crc kubenswrapper[4678]: E1013 12:45:37.592460 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:37 crc kubenswrapper[4678]: E1013 12:45:37.592549 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:37 crc kubenswrapper[4678]: E1013 12:45:37.706492 4678 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 13 12:45:38 crc kubenswrapper[4678]: I1013 12:45:38.591664 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:38 crc kubenswrapper[4678]: E1013 12:45:38.592074 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:39 crc kubenswrapper[4678]: I1013 12:45:39.591595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:39 crc kubenswrapper[4678]: I1013 12:45:39.591677 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:39 crc kubenswrapper[4678]: I1013 12:45:39.591619 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:39 crc kubenswrapper[4678]: E1013 12:45:39.591813 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:39 crc kubenswrapper[4678]: E1013 12:45:39.591942 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:39 crc kubenswrapper[4678]: E1013 12:45:39.592350 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:40 crc kubenswrapper[4678]: I1013 12:45:40.591771 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:40 crc kubenswrapper[4678]: E1013 12:45:40.592031 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:41 crc kubenswrapper[4678]: I1013 12:45:41.591986 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:41 crc kubenswrapper[4678]: I1013 12:45:41.592099 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:41 crc kubenswrapper[4678]: I1013 12:45:41.592108 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:41 crc kubenswrapper[4678]: E1013 12:45:41.592223 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:41 crc kubenswrapper[4678]: E1013 12:45:41.592400 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:41 crc kubenswrapper[4678]: E1013 12:45:41.592479 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:41 crc kubenswrapper[4678]: I1013 12:45:41.592931 4678 scope.go:117] "RemoveContainer" containerID="1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c" Oct 13 12:45:42 crc kubenswrapper[4678]: I1013 12:45:42.271496 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/1.log" Oct 13 12:45:42 crc kubenswrapper[4678]: I1013 12:45:42.271900 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerStarted","Data":"94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb"} Oct 13 12:45:42 crc kubenswrapper[4678]: I1013 12:45:42.592289 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:42 crc kubenswrapper[4678]: E1013 12:45:42.594247 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:42 crc kubenswrapper[4678]: E1013 12:45:42.707090 4678 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 13 12:45:43 crc kubenswrapper[4678]: I1013 12:45:43.592395 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:43 crc kubenswrapper[4678]: I1013 12:45:43.592452 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:43 crc kubenswrapper[4678]: I1013 12:45:43.592399 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:43 crc kubenswrapper[4678]: E1013 12:45:43.592565 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:43 crc kubenswrapper[4678]: E1013 12:45:43.592713 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:43 crc kubenswrapper[4678]: E1013 12:45:43.592842 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:44 crc kubenswrapper[4678]: I1013 12:45:44.592330 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:44 crc kubenswrapper[4678]: E1013 12:45:44.592555 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:45 crc kubenswrapper[4678]: I1013 12:45:45.592221 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:45 crc kubenswrapper[4678]: I1013 12:45:45.592314 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:45 crc kubenswrapper[4678]: E1013 12:45:45.592432 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:45 crc kubenswrapper[4678]: I1013 12:45:45.592446 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:45 crc kubenswrapper[4678]: E1013 12:45:45.592572 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:45 crc kubenswrapper[4678]: E1013 12:45:45.592702 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:46 crc kubenswrapper[4678]: I1013 12:45:46.592045 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:46 crc kubenswrapper[4678]: E1013 12:45:46.592325 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbs76" podUID="4ccdd1a7-2f11-43fd-9369-5ca8380ef70a" Oct 13 12:45:47 crc kubenswrapper[4678]: I1013 12:45:47.592355 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:47 crc kubenswrapper[4678]: I1013 12:45:47.592503 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:47 crc kubenswrapper[4678]: E1013 12:45:47.592540 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 13 12:45:47 crc kubenswrapper[4678]: I1013 12:45:47.592378 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:47 crc kubenswrapper[4678]: E1013 12:45:47.592728 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 13 12:45:47 crc kubenswrapper[4678]: E1013 12:45:47.592889 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 13 12:45:48 crc kubenswrapper[4678]: I1013 12:45:48.592543 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:45:48 crc kubenswrapper[4678]: I1013 12:45:48.596095 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 13 12:45:48 crc kubenswrapper[4678]: I1013 12:45:48.596318 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.592208 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.592262 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.592236 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.601783 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.601796 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.602023 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 13 12:45:49 crc kubenswrapper[4678]: I1013 12:45:49.602255 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.817322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.856944 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-shbr8"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.858207 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.859379 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.860305 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.860504 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dkxs9"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.861709 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.863048 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.863446 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.864089 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.864499 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.864717 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.865765 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5hccp"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.866232 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.867472 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.868113 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.870424 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.871894 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.874207 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.875137 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887476 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-client\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887564 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v82hd\" (UniqueName: \"kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887618 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887665 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt8qd\" (UniqueName: \"kubernetes.io/projected/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-kube-api-access-wt8qd\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7df0cb0-1905-4ca3-a888-82b7841582cd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887736 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmslb\" (UniqueName: \"kubernetes.io/projected/b7df0cb0-1905-4ca3-a888-82b7841582cd-kube-api-access-tmslb\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-serving-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887802 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-encryption-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887833 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-policies\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887864 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-image-import-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887912 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887944 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-images\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.887991 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-trusted-ca-bundle\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-node-pullsecrets\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888130 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75b4b6f2-c9a1-47d8-a8db-310811c8056a-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-serving-cert\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888224 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888259 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888291 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888331 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/75b4b6f2-c9a1-47d8-a8db-310811c8056a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888363 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7df0cb0-1905-4ca3-a888-82b7841582cd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888393 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit-dir\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888429 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-auth-proxy-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888493 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888526 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888561 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhvf2\" (UniqueName: \"kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888595 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888629 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crbcn\" (UniqueName: \"kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3da9a359-7295-4aa0-95e0-070be29e5eb9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888694 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnj9t\" (UniqueName: \"kubernetes.io/projected/3da9a359-7295-4aa0-95e0-070be29e5eb9-kube-api-access-dnj9t\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888775 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-config\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888812 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888856 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888888 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-encryption-config\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888919 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888952 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-serving-cert\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.888984 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj7nn\" (UniqueName: \"kubernetes.io/projected/75b4b6f2-c9a1-47d8-a8db-310811c8056a-kube-api-access-nj7nn\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889048 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9zdh\" (UniqueName: \"kubernetes.io/projected/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-kube-api-access-b9zdh\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889111 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889142 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpgj7\" (UniqueName: \"kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889173 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-dir\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889205 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-client\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889238 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.889270 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.890144 4678 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj": failed to list *v1.Secret: secrets "authentication-operator-dockercfg-mz9bj" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.890187 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-mz9bj\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"authentication-operator-dockercfg-mz9bj\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.890531 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.890535 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.890862 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.892234 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.893582 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.894396 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.894634 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.894770 4678 reflector.go:561] object-"openshift-authentication-operator"/"service-ca-bundle": failed to list *v1.ConfigMap: configmaps "service-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.894804 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"service-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"service-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.894877 4678 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4": failed to list *v1.Secret: secrets "machine-approver-sa-dockercfg-nl2j4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.894898 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-nl2j4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-sa-dockercfg-nl2j4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.894981 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.895190 4678 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-tls": failed to list *v1.Secret: secrets "machine-approver-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.895219 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.895296 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.895430 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.895789 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.895938 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.896117 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.896614 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.897324 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.897673 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.897952 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.898308 4678 reflector.go:561] object-"openshift-authentication-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898345 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.898362 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.898417 4678 reflector.go:561] object-"openshift-authentication-operator"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898445 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.898463 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.898428 4678 reflector.go:561] object-"openshift-authentication-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898486 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.898495 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898554 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898627 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898645 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898346 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898754 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.898809 4678 reflector.go:561] object-"openshift-authentication-operator"/"authentication-operator-config": failed to list *v1.ConfigMap: configmaps "authentication-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication-operator": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898822 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.898826 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication-operator\"/\"authentication-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"authentication-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898857 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.898960 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.897970 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899402 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899600 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899688 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899736 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899796 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 13 12:45:51 crc kubenswrapper[4678]: W1013 12:45:51.899879 4678 reflector.go:561] object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Oct 13 12:45:51 crc kubenswrapper[4678]: E1013 12:45:51.899906 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.899958 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900006 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900106 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900196 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900210 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900233 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900276 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900358 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.900908 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.901353 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.901531 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.901221 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.901237 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.903131 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qnj4p"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.903608 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.903989 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.904556 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.904647 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-c9rl2"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.904871 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.905251 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.905462 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.905724 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.905732 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.943787 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.944193 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.944310 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.944706 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.947599 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.947875 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.947903 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948022 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948083 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948098 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948190 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948542 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948858 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.948914 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.949825 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.950638 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.952966 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.953200 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.954627 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.954972 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.955280 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.958182 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.958668 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.958836 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.958948 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.960376 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.961474 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm2j9"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.961022 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.963415 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ldbrg"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.964011 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.965107 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.965620 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.965747 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v7vb4"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.965623 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.966920 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970338 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970607 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970476 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970515 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970514 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.970686 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.972247 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 13 12:45:51 crc kubenswrapper[4678]: I1013 12:45:51.973146 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.000178 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.000408 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.000704 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.000914 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:51.973189 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:51.973745 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:51.974766 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002157 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002219 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002751 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-tsjvf"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002824 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7df0cb0-1905-4ca3-a888-82b7841582cd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002873 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002897 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf080121-ca40-4473-9ca6-a8feb0780bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002918 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmslb\" (UniqueName: \"kubernetes.io/projected/b7df0cb0-1905-4ca3-a888-82b7841582cd-kube-api-access-tmslb\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002973 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-policies\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.002993 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-serving-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003013 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-encryption-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003036 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003068 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-image-import-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003109 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-images\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003130 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003150 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003171 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-node-pullsecrets\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003190 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-trusted-ca-bundle\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003206 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75b4b6f2-c9a1-47d8-a8db-310811c8056a-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-serving-cert\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003284 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003303 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003322 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003354 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003377 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003399 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003445 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/75b4b6f2-c9a1-47d8-a8db-310811c8056a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003465 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7df0cb0-1905-4ca3-a888-82b7841582cd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003484 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit-dir\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003507 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003525 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003550 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-auth-proxy-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003570 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003590 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003608 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003629 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003648 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003668 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003689 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhvf2\" (UniqueName: \"kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003730 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003750 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnj9t\" (UniqueName: \"kubernetes.io/projected/3da9a359-7295-4aa0-95e0-070be29e5eb9-kube-api-access-dnj9t\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crbcn\" (UniqueName: \"kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003791 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3da9a359-7295-4aa0-95e0-070be29e5eb9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003816 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003839 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003859 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh4z8\" (UniqueName: \"kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003891 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-config\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003912 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003933 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t66pk\" (UniqueName: \"kubernetes.io/projected/cf080121-ca40-4473-9ca6-a8feb0780bf2-kube-api-access-t66pk\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003930 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003970 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.003994 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004016 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-encryption-config\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004039 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-serving-cert\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004072 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj7nn\" (UniqueName: \"kubernetes.io/projected/75b4b6f2-c9a1-47d8-a8db-310811c8056a-kube-api-access-nj7nn\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004095 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004116 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9zdh\" (UniqueName: \"kubernetes.io/projected/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-kube-api-access-b9zdh\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004176 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpgj7\" (UniqueName: \"kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004194 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004202 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-dir\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004236 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-client\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004281 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004312 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-config\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004333 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt8qd\" (UniqueName: \"kubernetes.io/projected/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-kube-api-access-wt8qd\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004352 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-client\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004373 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v82hd\" (UniqueName: \"kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004402 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.005041 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7df0cb0-1905-4ca3-a888-82b7841582cd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.005159 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.005516 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-auth-proxy-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.005681 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.005911 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.006034 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b9796cc-d9c6-445d-8dce-1c048c21603e-config\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.006521 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.006772 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004208 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007103 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dkxs9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.012376 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007202 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007171 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.004378 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:51.974886 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007383 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007583 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.007923 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.009556 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.008584 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-policies\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.025883 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.026455 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.027117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-encryption-config\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.027557 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-serving-cert\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.028122 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-serving-cert\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.028203 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.028311 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029379 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029461 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-node-pullsecrets\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029485 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit-dir\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029497 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-audit-dir\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029678 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.029765 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/75b4b6f2-c9a1-47d8-a8db-310811c8056a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.030251 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5hccp"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.030269 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.030550 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.030876 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.030974 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.031014 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.032592 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-image-import-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.032888 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033027 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-config\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033472 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-serving-ca\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033532 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3da9a359-7295-4aa0-95e0-070be29e5eb9-images\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033688 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033937 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-trusted-ca-bundle\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.033953 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-audit\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.034197 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.037150 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-shbr8"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.038419 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.038494 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.038972 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nggsb"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.039372 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.039807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.034424 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.040039 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.040223 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.040353 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.036806 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.040569 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.041170 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.041836 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.044177 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.044693 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-65hzw"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.045025 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.045204 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.046554 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.046677 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.047242 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.048696 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.049314 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-encryption-config\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.051151 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vjqg7"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.051718 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.052316 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7df0cb0-1905-4ca3-a888-82b7841582cd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.054349 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.054469 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.054546 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ldbrg"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.054615 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.054737 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.055113 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.057806 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.058025 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.060546 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.064192 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.064613 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-etcd-client\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.065140 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.065178 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.065622 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3da9a359-7295-4aa0-95e0-070be29e5eb9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.066491 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.070131 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.073751 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75b4b6f2-c9a1-47d8-a8db-310811c8056a-serving-cert\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.075129 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.075281 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c9rl2"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.077477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-etcd-client\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.080101 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.082659 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm2j9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.082943 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.088462 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.092563 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.092788 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.102773 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qnj4p"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.105742 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hkh7k"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.106020 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-config\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108335 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf080121-ca40-4473-9ca6-a8feb0780bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108454 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108485 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108567 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108602 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108622 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108699 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108725 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108745 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108763 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108868 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108898 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.108964 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh4z8\" (UniqueName: \"kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.109045 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t66pk\" (UniqueName: \"kubernetes.io/projected/cf080121-ca40-4473-9ca6-a8feb0780bf2-kube-api-access-t66pk\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.109315 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-config\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.110989 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.112722 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.111647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114089 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114167 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114702 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114812 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bst8g"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114897 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.114969 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.115024 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.115172 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.115416 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.116441 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.116538 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.117623 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.117848 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.118011 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.119253 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-65hzw"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.121600 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v7vb4"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.125653 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.128669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hkh7k"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.130839 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.134178 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nggsb"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.136280 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.138598 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.139806 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.140326 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.142066 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.143495 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.145332 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.147514 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.148714 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bst8g"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.150381 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.152622 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.156450 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vjqg7"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.157555 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hr4hw"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.158469 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.160041 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.181557 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.200648 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.220079 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.239905 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.260201 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.280113 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.299822 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.320599 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.332433 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf080121-ca40-4473-9ca6-a8feb0780bf2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.340241 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.360991 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.383643 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.400376 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.420503 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.439651 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.460419 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.481173 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.501031 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.553652 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmslb\" (UniqueName: \"kubernetes.io/projected/b7df0cb0-1905-4ca3-a888-82b7841582cd-kube-api-access-tmslb\") pod \"openshift-apiserver-operator-796bbdcf4f-hzdvq\" (UID: \"b7df0cb0-1905-4ca3-a888-82b7841582cd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.590454 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crbcn\" (UniqueName: \"kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn\") pod \"route-controller-manager-6576b87f9c-45q2h\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.611922 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9zdh\" (UniqueName: \"kubernetes.io/projected/17cb88f3-b5b5-4d92-94c8-fdcdb873c224-kube-api-access-b9zdh\") pod \"apiserver-76f77b778f-shbr8\" (UID: \"17cb88f3-b5b5-4d92-94c8-fdcdb873c224\") " pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.616298 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt8qd\" (UniqueName: \"kubernetes.io/projected/f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd-kube-api-access-wt8qd\") pod \"apiserver-7bbb656c7d-gr9dd\" (UID: \"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.636144 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnj9t\" (UniqueName: \"kubernetes.io/projected/3da9a359-7295-4aa0-95e0-070be29e5eb9-kube-api-access-dnj9t\") pod \"machine-api-operator-5694c8668f-dkxs9\" (UID: \"3da9a359-7295-4aa0-95e0-070be29e5eb9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.682012 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj7nn\" (UniqueName: \"kubernetes.io/projected/75b4b6f2-c9a1-47d8-a8db-310811c8056a-kube-api-access-nj7nn\") pod \"openshift-config-operator-7777fb866f-q8tf2\" (UID: \"75b4b6f2-c9a1-47d8-a8db-310811c8056a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.697843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v82hd\" (UniqueName: \"kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd\") pod \"controller-manager-879f6c89f-s8h75\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.705968 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.715892 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.719823 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.726549 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.741135 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.760618 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.781300 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.797369 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.804015 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.821584 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.841390 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.853526 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.860762 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.865516 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.881140 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.901167 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.902533 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.923483 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.944073 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.948409 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2"] Oct 13 12:45:52 crc kubenswrapper[4678]: I1013 12:45:52.980683 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.012074 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.022978 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.031111 4678 secret.go:188] Couldn't get secret openshift-cluster-machine-approver/machine-approver-tls: failed to sync secret cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.031190 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls podName:6b9796cc-d9c6-445d-8dce-1c048c21603e nodeName:}" failed. No retries permitted until 2025-10-13 12:45:53.531171744 +0000 UTC m=+141.615709628 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-approver-tls" (UniqueName: "kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls") pod "machine-approver-56656f9798-cf28k" (UID: "6b9796cc-d9c6-445d-8dce-1c048c21603e") : failed to sync secret cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.031127 4678 secret.go:188] Couldn't get secret openshift-authentication-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.031347 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert podName:e8b410b3-0ca3-419a-9dcb-188b954a2136 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:53.531314067 +0000 UTC m=+141.615851961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert") pod "authentication-operator-69f744f599-5hccp" (UID: "e8b410b3-0ca3-419a-9dcb-188b954a2136") : failed to sync secret cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.033539 4678 configmap.go:193] Couldn't get configMap openshift-authentication-operator/authentication-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.033653 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config podName:e8b410b3-0ca3-419a-9dcb-188b954a2136 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:53.533634739 +0000 UTC m=+141.618172613 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config") pod "authentication-operator-69f744f599-5hccp" (UID: "e8b410b3-0ca3-419a-9dcb-188b954a2136") : failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.033686 4678 configmap.go:193] Couldn't get configMap openshift-authentication-operator/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.033709 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle podName:e8b410b3-0ca3-419a-9dcb-188b954a2136 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:53.53370248 +0000 UTC m=+141.618240364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle") pod "authentication-operator-69f744f599-5hccp" (UID: "e8b410b3-0ca3-419a-9dcb-188b954a2136") : failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.041974 4678 request.go:700] Waited for 1.005971334s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-metrics-certs-default&limit=500&resourceVersion=0 Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.043951 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.050027 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-shbr8"] Oct 13 12:45:53 crc kubenswrapper[4678]: W1013 12:45:53.060588 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17cb88f3_b5b5_4d92_94c8_fdcdb873c224.slice/crio-0645cf7780771517588a36425d6147e735824873c0f528f83d6bd077341f0fb4 WatchSource:0}: Error finding container 0645cf7780771517588a36425d6147e735824873c0f528f83d6bd077341f0fb4: Status 404 returned error can't find the container with id 0645cf7780771517588a36425d6147e735824873c0f528f83d6bd077341f0fb4 Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.060621 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.087394 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.102646 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.119875 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.121695 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd"] Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.139614 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.161308 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.173040 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq"] Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.177205 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.181785 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.200390 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.226168 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 13 12:45:53 crc kubenswrapper[4678]: W1013 12:45:53.233932 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e814849_60c6_4107_8ed6_5301f9cda527.slice/crio-ed83ec74a888daf7fe582b713b88897972b35faf42a11c6e3bc91ed9d2270b4d WatchSource:0}: Error finding container ed83ec74a888daf7fe582b713b88897972b35faf42a11c6e3bc91ed9d2270b4d: Status 404 returned error can't find the container with id ed83ec74a888daf7fe582b713b88897972b35faf42a11c6e3bc91ed9d2270b4d Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.239955 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.260858 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.269171 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-dkxs9"] Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.270178 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:45:53 crc kubenswrapper[4678]: W1013 12:45:53.279983 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3da9a359_7295_4aa0_95e0_070be29e5eb9.slice/crio-6bca4ce2bfc2bde6d458aab2ef7a6a46139d1e49110238cb915917d89d813c4f WatchSource:0}: Error finding container 6bca4ce2bfc2bde6d458aab2ef7a6a46139d1e49110238cb915917d89d813c4f: Status 404 returned error can't find the container with id 6bca4ce2bfc2bde6d458aab2ef7a6a46139d1e49110238cb915917d89d813c4f Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.280383 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: W1013 12:45:53.280937 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83521d15_1151_43e4_bd62_8aca811169dc.slice/crio-1280e4b0eb3a0a05adf01f5be44757766d8e5f877727091d4950e30f2708b060 WatchSource:0}: Error finding container 1280e4b0eb3a0a05adf01f5be44757766d8e5f877727091d4950e30f2708b060: Status 404 returned error can't find the container with id 1280e4b0eb3a0a05adf01f5be44757766d8e5f877727091d4950e30f2708b060 Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.300044 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.316863 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" event={"ID":"83521d15-1151-43e4-bd62-8aca811169dc","Type":"ContainerStarted","Data":"1280e4b0eb3a0a05adf01f5be44757766d8e5f877727091d4950e30f2708b060"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.318273 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" event={"ID":"17cb88f3-b5b5-4d92-94c8-fdcdb873c224","Type":"ContainerStarted","Data":"62c2fd4d26b493a9ae72368e6415e90bf54795d0758d70415ee266637e589d0a"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.318296 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" event={"ID":"17cb88f3-b5b5-4d92-94c8-fdcdb873c224","Type":"ContainerStarted","Data":"0645cf7780771517588a36425d6147e735824873c0f528f83d6bd077341f0fb4"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.319514 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" event={"ID":"3da9a359-7295-4aa0-95e0-070be29e5eb9","Type":"ContainerStarted","Data":"6bca4ce2bfc2bde6d458aab2ef7a6a46139d1e49110238cb915917d89d813c4f"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.320316 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.321162 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" event={"ID":"b7df0cb0-1905-4ca3-a888-82b7841582cd","Type":"ContainerStarted","Data":"d4b22153f01c2e57d0396e729d64145de2b3693ea2b84ce1b92efb85399c2d71"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.322486 4678 generic.go:334] "Generic (PLEG): container finished" podID="75b4b6f2-c9a1-47d8-a8db-310811c8056a" containerID="df7b5c91cc76f9e8e6b7d315d06240cfb2df6f90bb07a262bad1655183988645" exitCode=0 Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.322545 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" event={"ID":"75b4b6f2-c9a1-47d8-a8db-310811c8056a","Type":"ContainerDied","Data":"df7b5c91cc76f9e8e6b7d315d06240cfb2df6f90bb07a262bad1655183988645"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.322565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" event={"ID":"75b4b6f2-c9a1-47d8-a8db-310811c8056a","Type":"ContainerStarted","Data":"99b02c765a41dcf2e21d62ab31b17967f8fe16da81b6ddb1ced312c954d29f67"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.324524 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" event={"ID":"7e814849-60c6-4107-8ed6-5301f9cda527","Type":"ContainerStarted","Data":"ed83ec74a888daf7fe582b713b88897972b35faf42a11c6e3bc91ed9d2270b4d"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.325386 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" event={"ID":"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd","Type":"ContainerStarted","Data":"bcb66cd3b51acedbe06e63e24fef2a36a85b816c60c105db4699e50a30e9d65d"} Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.339990 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.360520 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.380581 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.399881 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.421209 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.441727 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.462990 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.480714 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.500323 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.520379 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.542136 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.561752 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.580671 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.600570 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.620213 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.627192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.627294 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.627361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.627518 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.641333 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.652457 4678 projected.go:288] Couldn't get configMap openshift-authentication-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.659694 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.680103 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.701068 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.719499 4678 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.719547 4678 projected.go:194] Error preparing data for projected volume kube-api-access-kpgj7 for pod openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: E1013 12:45:53.719635 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7 podName:6b9796cc-d9c6-445d-8dce-1c048c21603e nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.219604313 +0000 UTC m=+142.304142237 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-kpgj7" (UniqueName: "kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7") pod "machine-approver-56656f9798-cf28k" (UID: "6b9796cc-d9c6-445d-8dce-1c048c21603e") : failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.721519 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.741287 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.760008 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.781011 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.800707 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.820352 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.840923 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.860874 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.881817 4678 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.900285 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.942616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t66pk\" (UniqueName: \"kubernetes.io/projected/cf080121-ca40-4473-9ca6-a8feb0780bf2-kube-api-access-t66pk\") pod \"multus-admission-controller-857f4d67dd-ldbrg\" (UID: \"cf080121-ca40-4473-9ca6-a8feb0780bf2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.959891 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2928d6d7-52d6-40ca-b351-3efbc1bfe6ff-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-8vwh8\" (UID: \"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.979112 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh4z8\" (UniqueName: \"kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8\") pod \"oauth-openshift-558db77b4-4ggx9\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.980925 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 13 12:45:53 crc kubenswrapper[4678]: I1013 12:45:53.992221 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.000096 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.019204 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.022174 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.041513 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.059026 4678 request.go:700] Waited for 1.942350046s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Ddefault-dockercfg-2llfx&limit=500&resourceVersion=0 Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.060447 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.081495 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.101745 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.120245 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.142300 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.161573 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.202903 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.206413 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8"] Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.207713 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-config\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.225408 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.235910 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.235945 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.235976 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.235999 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236040 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236070 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236088 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a497df52-2462-4299-b266-6f5ef91edaef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236104 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9nk2\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-kube-api-access-h9nk2\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfzrx\" (UniqueName: \"kubernetes.io/projected/c3d0f53a-d207-44be-a692-894d2740fa54-kube-api-access-dfzrx\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236177 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236193 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk2px\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-kube-api-access-lk2px\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236207 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-client\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236223 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-service-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236253 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236278 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ea4f437-7feb-4ca7-9e73-631c8dad195e-metrics-tls\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236301 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236343 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ea4f437-7feb-4ca7-9e73-631c8dad195e-trusted-ca\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236371 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236379 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ldbrg"] Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236395 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236425 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbf24aeb-11f0-4642-ac35-c69bd78242b5-metrics-tls\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236443 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b9d0828-76bc-451a-875a-bb2fb0f56322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236460 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236486 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67p86\" (UniqueName: \"kubernetes.io/projected/f35010e9-46ee-4ba5-8e85-b7568bf88650-kube-api-access-67p86\") pod \"downloads-7954f5f757-c9rl2\" (UID: \"f35010e9-46ee-4ba5-8e85-b7568bf88650\") " pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236505 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236520 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b9d0828-76bc-451a-875a-bb2fb0f56322-proxy-tls\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236537 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236553 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236567 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnbsr\" (UniqueName: \"kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236582 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mbkr\" (UniqueName: \"kubernetes.io/projected/48f355d2-c241-44aa-a567-e9c31de5d18e-kube-api-access-8mbkr\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236597 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bljtq\" (UniqueName: \"kubernetes.io/projected/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-kube-api-access-bljtq\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236612 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-config\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236647 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-config\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a497df52-2462-4299-b266-6f5ef91edaef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.236670 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.736658765 +0000 UTC m=+142.821196649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236693 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236716 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236734 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-config\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236790 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f355d2-c241-44aa-a567-e9c31de5d18e-serving-cert\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236807 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-trusted-ca\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236837 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3d0f53a-d207-44be-a692-894d2740fa54-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236852 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-serving-cert\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236891 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdd4\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236907 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwjk8\" (UniqueName: \"kubernetes.io/projected/7b9d0828-76bc-451a-875a-bb2fb0f56322-kube-api-access-bwjk8\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236922 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6hjl\" (UniqueName: \"kubernetes.io/projected/96d06121-ad51-4e70-9a3a-952c6863efed-kube-api-access-n6hjl\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236938 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrrqz\" (UniqueName: \"kubernetes.io/projected/dbf24aeb-11f0-4642-ac35-c69bd78242b5-kube-api-access-qrrqz\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.236956 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.237019 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpgj7\" (UniqueName: \"kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.237036 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3d0f53a-d207-44be-a692-894d2740fa54-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.237066 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.237082 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.239857 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 13 12:45:54 crc kubenswrapper[4678]: W1013 12:45:54.249079 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf080121_ca40_4473_9ca6_a8feb0780bf2.slice/crio-3cffd283235d7b3c5aaa12bede48daa27d5bb4c653e934683ce34d1511475acc WatchSource:0}: Error finding container 3cffd283235d7b3c5aaa12bede48daa27d5bb4c653e934683ce34d1511475acc: Status 404 returned error can't find the container with id 3cffd283235d7b3c5aaa12bede48daa27d5bb4c653e934683ce34d1511475acc Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.256112 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8b410b3-0ca3-419a-9dcb-188b954a2136-serving-cert\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.260808 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.270772 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b9796cc-d9c6-445d-8dce-1c048c21603e-machine-approver-tls\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.300948 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.326514 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.329494 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e8b410b3-0ca3-419a-9dcb-188b954a2136-service-ca-bundle\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.337732 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.337990 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80adeb99-a848-4261-a789-70cd9c333430-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338020 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrsxp\" (UniqueName: \"kubernetes.io/projected/80adeb99-a848-4261-a789-70cd9c333430-kube-api-access-xrsxp\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338044 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338209 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338237 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad4e211d-f191-4262-9e70-d1f4fcaddeff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338282 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk2px\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-kube-api-access-lk2px\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338305 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-service-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338327 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21bd5a7f-366f-4d05-986d-6285b95bb1a4-config-volume\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338351 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/99ba1b17-157b-4df8-b753-52b5742443cd-tmpfs\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338399 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ea4f437-7feb-4ca7-9e73-631c8dad195e-metrics-tls\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338442 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bee92f-4cf8-476d-8d83-8eddce9a34ab-cert\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338508 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338529 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5632c50a-0b60-417c-afc1-633d42baf685-serving-cert\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338551 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hqxq\" (UniqueName: \"kubernetes.io/projected/6e30bbf0-723d-4233-b05f-a679c0e32ad5-kube-api-access-2hqxq\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338574 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b79db0ac-d465-4200-acc8-194524524a3a-proxy-tls\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338594 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-certs\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338636 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-metrics-certs\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338667 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfj8s\" (UniqueName: \"kubernetes.io/projected/916d3614-f8db-4446-bda3-224c9bc15357-kube-api-access-rfj8s\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338690 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b9d0828-76bc-451a-875a-bb2fb0f56322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338723 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-apiservice-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67p86\" (UniqueName: \"kubernetes.io/projected/f35010e9-46ee-4ba5-8e85-b7568bf88650-kube-api-access-67p86\") pod \"downloads-7954f5f757-c9rl2\" (UID: \"f35010e9-46ee-4ba5-8e85-b7568bf88650\") " pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338798 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsvzh\" (UniqueName: \"kubernetes.io/projected/80104b9e-1a05-40d6-b179-f9330679357a-kube-api-access-vsvzh\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338823 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338850 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b9d0828-76bc-451a-875a-bb2fb0f56322-proxy-tls\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338873 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-stats-auth\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338899 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338924 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnbsr\" (UniqueName: \"kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338948 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mbkr\" (UniqueName: \"kubernetes.io/projected/48f355d2-c241-44aa-a567-e9c31de5d18e-kube-api-access-8mbkr\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.338973 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bljtq\" (UniqueName: \"kubernetes.io/projected/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-kube-api-access-bljtq\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339002 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r947\" (UniqueName: \"kubernetes.io/projected/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-kube-api-access-7r947\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339075 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-config\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.339109 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.8390887 +0000 UTC m=+142.923626584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339177 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-config\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339199 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a497df52-2462-4299-b266-6f5ef91edaef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339217 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339233 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-config\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339255 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e30bbf0-723d-4233-b05f-a679c0e32ad5-service-ca-bundle\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339274 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6dc2394-2372-4e43-ae9f-8557a979464c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-trusted-ca\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw2cd\" (UniqueName: \"kubernetes.io/projected/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-kube-api-access-dw2cd\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339353 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwjk8\" (UniqueName: \"kubernetes.io/projected/7b9d0828-76bc-451a-875a-bb2fb0f56322-kube-api-access-bwjk8\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339375 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339395 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-node-bootstrap-token\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klwtj\" (UniqueName: \"kubernetes.io/projected/02bee92f-4cf8-476d-8d83-8eddce9a34ab-kube-api-access-klwtj\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339439 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48wrb\" (UniqueName: \"kubernetes.io/projected/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-kube-api-access-48wrb\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3d0f53a-d207-44be-a692-894d2740fa54-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339489 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339504 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339529 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-default-certificate\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339546 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339562 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-csi-data-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339576 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-srv-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339599 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339632 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339648 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a497df52-2462-4299-b266-6f5ef91edaef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339664 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9nk2\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-kube-api-access-h9nk2\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339679 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-registration-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339697 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-client\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfzrx\" (UniqueName: \"kubernetes.io/projected/c3d0f53a-d207-44be-a692-894d2740fa54-kube-api-access-dfzrx\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339752 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-profile-collector-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339768 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339791 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339807 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-plugins-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339823 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339839 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5632c50a-0b60-417c-afc1-633d42baf685-config\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339876 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ea4f437-7feb-4ca7-9e73-631c8dad195e-trusted-ca\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339921 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lt2\" (UniqueName: \"kubernetes.io/projected/098d2671-5613-4906-af73-dc25b2ebd868-kube-api-access-t8lt2\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339942 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-config\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.339951 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80adeb99-a848-4261-a789-70cd9c333430-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340033 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340095 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340182 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq2n5\" (UniqueName: \"kubernetes.io/projected/65ee6db4-4e7e-41e8-9a6f-14c54e13f36b-kube-api-access-jq2n5\") pod \"migrator-59844c95c7-6bgtc\" (UID: \"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340211 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w956\" (UniqueName: \"kubernetes.io/projected/b79db0ac-d465-4200-acc8-194524524a3a-kube-api-access-7w956\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbf24aeb-11f0-4642-ac35-c69bd78242b5-metrics-tls\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340261 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-srv-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340292 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6dc2394-2372-4e43-ae9f-8557a979464c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340323 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sd69\" (UniqueName: \"kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340352 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6dc2394-2372-4e43-ae9f-8557a979464c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340381 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340407 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh7pr\" (UniqueName: \"kubernetes.io/projected/ad4e211d-f191-4262-9e70-d1f4fcaddeff-kube-api-access-vh7pr\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340452 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340478 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/80104b9e-1a05-40d6-b179-f9330679357a-signing-key\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340499 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgnf\" (UniqueName: \"kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340548 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-images\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340587 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340611 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg9t2\" (UniqueName: \"kubernetes.io/projected/99ba1b17-157b-4df8-b753-52b5742443cd-kube-api-access-dg9t2\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340634 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56tts\" (UniqueName: \"kubernetes.io/projected/5632c50a-0b60-417c-afc1-633d42baf685-kube-api-access-56tts\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340691 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f355d2-c241-44aa-a567-e9c31de5d18e-serving-cert\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340716 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-webhook-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340756 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3d0f53a-d207-44be-a692-894d2740fa54-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-config\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-serving-cert\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.340957 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-mountpoint-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.341024 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdd4\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.341144 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrrqz\" (UniqueName: \"kubernetes.io/projected/dbf24aeb-11f0-4642-ac35-c69bd78242b5-kube-api-access-qrrqz\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.341175 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6hjl\" (UniqueName: \"kubernetes.io/projected/96d06121-ad51-4e70-9a3a-952c6863efed-kube-api-access-n6hjl\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.341203 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-socket-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.341226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21bd5a7f-366f-4d05-986d-6285b95bb1a4-metrics-tls\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.346279 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/80104b9e-1a05-40d6-b179-f9330679357a-signing-cabundle\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.346406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.346450 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-984fr\" (UniqueName: \"kubernetes.io/projected/21bd5a7f-366f-4d05-986d-6285b95bb1a4-kube-api-access-984fr\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.348817 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a497df52-2462-4299-b266-6f5ef91edaef-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.349246 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.352088 4678 projected.go:194] Error preparing data for projected volume kube-api-access-zhvf2 for pod openshift-authentication-operator/authentication-operator-69f744f599-5hccp: failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.353221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-config\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.354218 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/48f355d2-c241-44aa-a567-e9c31de5d18e-trusted-ca\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.355374 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.356612 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" event={"ID":"cf080121-ca40-4473-9ca6-a8feb0780bf2","Type":"ContainerStarted","Data":"3cffd283235d7b3c5aaa12bede48daa27d5bb4c653e934683ce34d1511475acc"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.357037 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.357213 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a497df52-2462-4299-b266-6f5ef91edaef-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.357448 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3d0f53a-d207-44be-a692-894d2740fa54-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.357565 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ea4f437-7feb-4ca7-9e73-631c8dad195e-trusted-ca\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.358020 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.358101 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.358470 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.858450099 +0000 UTC m=+142.942987973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.359472 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2 podName:e8b410b3-0ca3-419a-9dcb-188b954a2136 nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.859456706 +0000 UTC m=+142.943994590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-zhvf2" (UniqueName: "kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2") pod "authentication-operator-69f744f599-5hccp" (UID: "e8b410b3-0ca3-419a-9dcb-188b954a2136") : failed to sync configmap cache: timed out waiting for the condition Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.359483 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.360549 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-service-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.361290 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-ca\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.361781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.362314 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b9d0828-76bc-451a-875a-bb2fb0f56322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.362870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.363587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.363755 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.364085 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-etcd-client\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.364887 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96d06121-ad51-4e70-9a3a-952c6863efed-serving-cert\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.365034 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.365090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" event={"ID":"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff","Type":"ContainerStarted","Data":"e88cda36718771d3cde9b5c7d57fb6ddabe91f2ba35b81c4f32dd60e24080a29"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.365483 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.365512 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" event={"ID":"b7df0cb0-1905-4ca3-a888-82b7841582cd","Type":"ContainerStarted","Data":"587649e61363f1cb3fc60b03e24406ebc09284cad031cddc84900e2432006a78"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.365520 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/dbf24aeb-11f0-4642-ac35-c69bd78242b5-metrics-tls\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.366441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.367380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5ea4f437-7feb-4ca7-9e73-631c8dad195e-metrics-tls\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.368967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f355d2-c241-44aa-a567-e9c31de5d18e-serving-cert\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.369540 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c3d0f53a-d207-44be-a692-894d2740fa54-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.370602 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.371223 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.375631 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpgj7\" (UniqueName: \"kubernetes.io/projected/6b9796cc-d9c6-445d-8dce-1c048c21603e-kube-api-access-kpgj7\") pod \"machine-approver-56656f9798-cf28k\" (UID: \"6b9796cc-d9c6-445d-8dce-1c048c21603e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.380428 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.381064 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" event={"ID":"83521d15-1151-43e4-bd62-8aca811169dc","Type":"ContainerStarted","Data":"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.381531 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.382589 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b9d0828-76bc-451a-875a-bb2fb0f56322-proxy-tls\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.383689 4678 generic.go:334] "Generic (PLEG): container finished" podID="17cb88f3-b5b5-4d92-94c8-fdcdb873c224" containerID="62c2fd4d26b493a9ae72368e6415e90bf54795d0758d70415ee266637e589d0a" exitCode=0 Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.383715 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" event={"ID":"17cb88f3-b5b5-4d92-94c8-fdcdb873c224","Type":"ContainerDied","Data":"62c2fd4d26b493a9ae72368e6415e90bf54795d0758d70415ee266637e589d0a"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.386148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" event={"ID":"3da9a359-7295-4aa0-95e0-070be29e5eb9","Type":"ContainerStarted","Data":"e1baf661c5bcb214e5d6bd466c6dd7e1165e2a140155b71a08772d879e62b4d8"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.386173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" event={"ID":"3da9a359-7295-4aa0-95e0-070be29e5eb9","Type":"ContainerStarted","Data":"556c9235bb7711be112737c1714137694e75589c9dfa7c9a064c17058f0ccc02"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.389714 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.390547 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" event={"ID":"7e814849-60c6-4107-8ed6-5301f9cda527","Type":"ContainerStarted","Data":"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.390815 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.394209 4678 generic.go:334] "Generic (PLEG): container finished" podID="f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd" containerID="a2e7d4af7a30fadf1e0a0cfddbad9516e1cf5a73410c88b47fd87e39458c79c7" exitCode=0 Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.394258 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" event={"ID":"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd","Type":"ContainerDied","Data":"a2e7d4af7a30fadf1e0a0cfddbad9516e1cf5a73410c88b47fd87e39458c79c7"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.394878 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.403616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" event={"ID":"75b4b6f2-c9a1-47d8-a8db-310811c8056a","Type":"ContainerStarted","Data":"cb93c4429d3bb83670b017b24185b73a24994ac9b7570295db3baeddd2ddcf37"} Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.404284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.444107 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.446477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447394 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447570 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/80104b9e-1a05-40d6-b179-f9330679357a-signing-key\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.447624 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.947597464 +0000 UTC m=+143.032135348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447664 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgnf\" (UniqueName: \"kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-images\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447742 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg9t2\" (UniqueName: \"kubernetes.io/projected/99ba1b17-157b-4df8-b753-52b5742443cd-kube-api-access-dg9t2\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447769 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56tts\" (UniqueName: \"kubernetes.io/projected/5632c50a-0b60-417c-afc1-633d42baf685-kube-api-access-56tts\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447834 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-webhook-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447876 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-mountpoint-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447964 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-socket-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.447983 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21bd5a7f-366f-4d05-986d-6285b95bb1a4-metrics-tls\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448000 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/80104b9e-1a05-40d6-b179-f9330679357a-signing-cabundle\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448028 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448045 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-984fr\" (UniqueName: \"kubernetes.io/projected/21bd5a7f-366f-4d05-986d-6285b95bb1a4-kube-api-access-984fr\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448104 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80adeb99-a848-4261-a789-70cd9c333430-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448126 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrsxp\" (UniqueName: \"kubernetes.io/projected/80adeb99-a848-4261-a789-70cd9c333430-kube-api-access-xrsxp\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad4e211d-f191-4262-9e70-d1f4fcaddeff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448193 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21bd5a7f-366f-4d05-986d-6285b95bb1a4-config-volume\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448225 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/99ba1b17-157b-4df8-b753-52b5742443cd-tmpfs\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448279 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bee92f-4cf8-476d-8d83-8eddce9a34ab-cert\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448328 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hqxq\" (UniqueName: \"kubernetes.io/projected/6e30bbf0-723d-4233-b05f-a679c0e32ad5-kube-api-access-2hqxq\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448365 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5632c50a-0b60-417c-afc1-633d42baf685-serving-cert\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448382 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b79db0ac-d465-4200-acc8-194524524a3a-proxy-tls\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448414 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-certs\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448431 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfj8s\" (UniqueName: \"kubernetes.io/projected/916d3614-f8db-4446-bda3-224c9bc15357-kube-api-access-rfj8s\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-metrics-certs\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448480 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-apiservice-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448496 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsvzh\" (UniqueName: \"kubernetes.io/projected/80104b9e-1a05-40d6-b179-f9330679357a-kube-api-access-vsvzh\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448511 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448521 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448542 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-stats-auth\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448572 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r947\" (UniqueName: \"kubernetes.io/projected/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-kube-api-access-7r947\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448596 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-images\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448606 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e30bbf0-723d-4233-b05f-a679c0e32ad5-service-ca-bundle\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448623 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6dc2394-2372-4e43-ae9f-8557a979464c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448646 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw2cd\" (UniqueName: \"kubernetes.io/projected/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-kube-api-access-dw2cd\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-node-bootstrap-token\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448694 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48wrb\" (UniqueName: \"kubernetes.io/projected/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-kube-api-access-48wrb\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448718 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klwtj\" (UniqueName: \"kubernetes.io/projected/02bee92f-4cf8-476d-8d83-8eddce9a34ab-kube-api-access-klwtj\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448754 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-default-certificate\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448774 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-srv-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448792 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448810 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-csi-data-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448833 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-registration-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-profile-collector-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448887 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448908 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-plugins-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448946 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5632c50a-0b60-417c-afc1-633d42baf685-config\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448977 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lt2\" (UniqueName: \"kubernetes.io/projected/098d2671-5613-4906-af73-dc25b2ebd868-kube-api-access-t8lt2\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.448997 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80adeb99-a848-4261-a789-70cd9c333430-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449038 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq2n5\" (UniqueName: \"kubernetes.io/projected/65ee6db4-4e7e-41e8-9a6f-14c54e13f36b-kube-api-access-jq2n5\") pod \"migrator-59844c95c7-6bgtc\" (UID: \"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w956\" (UniqueName: \"kubernetes.io/projected/b79db0ac-d465-4200-acc8-194524524a3a-kube-api-access-7w956\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449194 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-srv-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6dc2394-2372-4e43-ae9f-8557a979464c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449255 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh7pr\" (UniqueName: \"kubernetes.io/projected/ad4e211d-f191-4262-9e70-d1f4fcaddeff-kube-api-access-vh7pr\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449271 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sd69\" (UniqueName: \"kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449286 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6dc2394-2372-4e43-ae9f-8557a979464c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.449863 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6dc2394-2372-4e43-ae9f-8557a979464c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.452454 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80adeb99-a848-4261-a789-70cd9c333430-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.452543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/21bd5a7f-366f-4d05-986d-6285b95bb1a4-metrics-tls\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.452549 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-default-certificate\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.452658 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-srv-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.453426 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/80104b9e-1a05-40d6-b179-f9330679357a-signing-key\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.453617 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:54.953606002 +0000 UTC m=+143.038143886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.453785 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-csi-data-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.453885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-registration-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.454781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80adeb99-a848-4261-a789-70cd9c333430-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.454875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-mountpoint-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.455029 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-socket-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.456001 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.456075 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b79db0ac-d465-4200-acc8-194524524a3a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.456901 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-webhook-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.456945 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-plugins-dir\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.458619 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5632c50a-0b60-417c-afc1-633d42baf685-config\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.458637 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/80104b9e-1a05-40d6-b179-f9330679357a-signing-cabundle\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.460126 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-w87mp\" (UID: \"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.462350 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/02bee92f-4cf8-476d-8d83-8eddce9a34ab-cert\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.462400 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.462887 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.463184 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-profile-collector-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.463416 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6e30bbf0-723d-4233-b05f-a679c0e32ad5-service-ca-bundle\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.463457 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6dc2394-2372-4e43-ae9f-8557a979464c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.463691 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/99ba1b17-157b-4df8-b753-52b5742443cd-tmpfs\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.464892 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21bd5a7f-366f-4d05-986d-6285b95bb1a4-config-volume\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.467139 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/99ba1b17-157b-4df8-b753-52b5742443cd-apiservice-cert\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.467156 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5632c50a-0b60-417c-afc1-633d42baf685-serving-cert\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.467360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-metrics-certs\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.468366 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/6e30bbf0-723d-4233-b05f-a679c0e32ad5-stats-auth\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.468397 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad4e211d-f191-4262-9e70-d1f4fcaddeff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.468718 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b79db0ac-d465-4200-acc8-194524524a3a-proxy-tls\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.468867 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.469644 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-certs\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.471293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.474272 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/098d2671-5613-4906-af73-dc25b2ebd868-srv-cert\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.477076 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67p86\" (UniqueName: \"kubernetes.io/projected/f35010e9-46ee-4ba5-8e85-b7568bf88650-kube-api-access-67p86\") pod \"downloads-7954f5f757-c9rl2\" (UID: \"f35010e9-46ee-4ba5-8e85-b7568bf88650\") " pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.478206 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/916d3614-f8db-4446-bda3-224c9bc15357-node-bootstrap-token\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.480265 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.500243 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwjk8\" (UniqueName: \"kubernetes.io/projected/7b9d0828-76bc-451a-875a-bb2fb0f56322-kube-api-access-bwjk8\") pod \"machine-config-controller-84d6567774-8hr85\" (UID: \"7b9d0828-76bc-451a-875a-bb2fb0f56322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.530870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfzrx\" (UniqueName: \"kubernetes.io/projected/c3d0f53a-d207-44be-a692-894d2740fa54-kube-api-access-dfzrx\") pod \"openshift-controller-manager-operator-756b6f6bc6-wmbl2\" (UID: \"c3d0f53a-d207-44be-a692-894d2740fa54\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.543666 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.548792 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9nk2\" (UniqueName: \"kubernetes.io/projected/5ea4f437-7feb-4ca7-9e73-631c8dad195e-kube-api-access-h9nk2\") pod \"ingress-operator-5b745b69d9-rqc8w\" (UID: \"5ea4f437-7feb-4ca7-9e73-631c8dad195e\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.550292 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.554867 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.054844506 +0000 UTC m=+143.139382390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.555040 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.557078 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.057062504 +0000 UTC m=+143.141600388 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.557711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnbsr\" (UniqueName: \"kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr\") pod \"console-f9d7485db-qq6wd\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.573310 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.580747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mbkr\" (UniqueName: \"kubernetes.io/projected/48f355d2-c241-44aa-a567-e9c31de5d18e-kube-api-access-8mbkr\") pod \"console-operator-58897d9998-qnj4p\" (UID: \"48f355d2-c241-44aa-a567-e9c31de5d18e\") " pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.586129 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.593847 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bljtq\" (UniqueName: \"kubernetes.io/projected/214c4966-ab84-4ea0-acf4-8446bd0f2cd0-kube-api-access-bljtq\") pod \"cluster-samples-operator-665b6dd947-6dcl8\" (UID: \"214c4966-ab84-4ea0-acf4-8446bd0f2cd0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.598342 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.612603 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.615347 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk2px\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-kube-api-access-lk2px\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.642491 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.646102 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.658933 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.659322 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.159300634 +0000 UTC m=+143.243838518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.667121 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrrqz\" (UniqueName: \"kubernetes.io/projected/dbf24aeb-11f0-4642-ac35-c69bd78242b5-kube-api-access-qrrqz\") pod \"dns-operator-744455d44c-jm2j9\" (UID: \"dbf24aeb-11f0-4642-ac35-c69bd78242b5\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.682268 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6hjl\" (UniqueName: \"kubernetes.io/projected/96d06121-ad51-4e70-9a3a-952c6863efed-kube-api-access-n6hjl\") pod \"etcd-operator-b45778765-v7vb4\" (UID: \"96d06121-ad51-4e70-9a3a-952c6863efed\") " pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.704996 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdd4\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.726544 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a497df52-2462-4299-b266-6f5ef91edaef-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-b5mc9\" (UID: \"a497df52-2462-4299-b266-6f5ef91edaef\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.758443 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgnf\" (UniqueName: \"kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf\") pod \"marketplace-operator-79b997595-pj7h9\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.765043 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.765619 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.26560357 +0000 UTC m=+143.350141454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.785713 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg9t2\" (UniqueName: \"kubernetes.io/projected/99ba1b17-157b-4df8-b753-52b5742443cd-kube-api-access-dg9t2\") pod \"packageserver-d55dfcdfc-9j7x8\" (UID: \"99ba1b17-157b-4df8-b753-52b5742443cd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.809259 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56tts\" (UniqueName: \"kubernetes.io/projected/5632c50a-0b60-417c-afc1-633d42baf685-kube-api-access-56tts\") pod \"service-ca-operator-777779d784-65hzw\" (UID: \"5632c50a-0b60-417c-afc1-633d42baf685\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.830811 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrsxp\" (UniqueName: \"kubernetes.io/projected/80adeb99-a848-4261-a789-70cd9c333430-kube-api-access-xrsxp\") pod \"kube-storage-version-migrator-operator-b67b599dd-jbw5p\" (UID: \"80adeb99-a848-4261-a789-70cd9c333430\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.852133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq2n5\" (UniqueName: \"kubernetes.io/projected/65ee6db4-4e7e-41e8-9a6f-14c54e13f36b-kube-api-access-jq2n5\") pod \"migrator-59844c95c7-6bgtc\" (UID: \"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.852390 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.863273 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.867398 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.867599 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.367583113 +0000 UTC m=+143.452120997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.867516 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.870745 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhvf2\" (UniqueName: \"kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.871177 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.870744 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lt2\" (UniqueName: \"kubernetes.io/projected/098d2671-5613-4906-af73-dc25b2ebd868-kube-api-access-t8lt2\") pod \"catalog-operator-68c6474976-xlrtj\" (UID: \"098d2671-5613-4906-af73-dc25b2ebd868\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.872119 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.372103342 +0000 UTC m=+143.456641226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.876086 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhvf2\" (UniqueName: \"kubernetes.io/projected/e8b410b3-0ca3-419a-9dcb-188b954a2136-kube-api-access-zhvf2\") pod \"authentication-operator-69f744f599-5hccp\" (UID: \"e8b410b3-0ca3-419a-9dcb-188b954a2136\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.877996 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh7pr\" (UniqueName: \"kubernetes.io/projected/ad4e211d-f191-4262-9e70-d1f4fcaddeff-kube-api-access-vh7pr\") pod \"control-plane-machine-set-operator-78cbb6b69f-m99jv\" (UID: \"ad4e211d-f191-4262-9e70-d1f4fcaddeff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.905638 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.923012 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w956\" (UniqueName: \"kubernetes.io/projected/b79db0ac-d465-4200-acc8-194524524a3a-kube-api-access-7w956\") pod \"machine-config-operator-74547568cd-8cp6h\" (UID: \"b79db0ac-d465-4200-acc8-194524524a3a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.940742 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.946597 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sd69\" (UniqueName: \"kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69\") pod \"collect-profiles-29339325-jl5mc\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.948825 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hqxq\" (UniqueName: \"kubernetes.io/projected/6e30bbf0-723d-4233-b05f-a679c0e32ad5-kube-api-access-2hqxq\") pod \"router-default-5444994796-tsjvf\" (UID: \"6e30bbf0-723d-4233-b05f-a679c0e32ad5\") " pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.959456 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.966004 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.971798 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.972509 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:54 crc kubenswrapper[4678]: E1013 12:45:54.973096 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.473076608 +0000 UTC m=+143.557614492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.975362 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d6dc2394-2372-4e43-ae9f-8557a979464c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cm2qp\" (UID: \"d6dc2394-2372-4e43-ae9f-8557a979464c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.978376 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsvzh\" (UniqueName: \"kubernetes.io/projected/80104b9e-1a05-40d6-b179-f9330679357a-kube-api-access-vsvzh\") pod \"service-ca-9c57cc56f-nggsb\" (UID: \"80104b9e-1a05-40d6-b179-f9330679357a\") " pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.980168 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.994581 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" Oct 13 12:45:54 crc kubenswrapper[4678]: I1013 12:45:54.995211 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-984fr\" (UniqueName: \"kubernetes.io/projected/21bd5a7f-366f-4d05-986d-6285b95bb1a4-kube-api-access-984fr\") pod \"dns-default-hkh7k\" (UID: \"21bd5a7f-366f-4d05-986d-6285b95bb1a4\") " pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.000302 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.008224 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.014360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.022701 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.034899 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r947\" (UniqueName: \"kubernetes.io/projected/1ba5a0dd-a844-42cf-80f0-1c49802e73e0-kube-api-access-7r947\") pod \"package-server-manager-789f6589d5-2slxb\" (UID: \"1ba5a0dd-a844-42cf-80f0-1c49802e73e0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.035152 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.059542 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48wrb\" (UniqueName: \"kubernetes.io/projected/3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1-kube-api-access-48wrb\") pod \"olm-operator-6b444d44fb-xbd8l\" (UID: \"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.061285 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw2cd\" (UniqueName: \"kubernetes.io/projected/56d1be59-35a1-44ac-873a-d5bc3df2f9ab-kube-api-access-dw2cd\") pod \"csi-hostpathplugin-vjqg7\" (UID: \"56d1be59-35a1-44ac-873a-d5bc3df2f9ab\") " pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.067638 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.074943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.075443 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.57542429 +0000 UTC m=+143.659962174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.088236 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfj8s\" (UniqueName: \"kubernetes.io/projected/916d3614-f8db-4446-bda3-224c9bc15357-kube-api-access-rfj8s\") pod \"machine-config-server-hr4hw\" (UID: \"916d3614-f8db-4446-bda3-224c9bc15357\") " pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.089963 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.097506 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klwtj\" (UniqueName: \"kubernetes.io/projected/02bee92f-4cf8-476d-8d83-8eddce9a34ab-kube-api-access-klwtj\") pod \"ingress-canary-bst8g\" (UID: \"02bee92f-4cf8-476d-8d83-8eddce9a34ab\") " pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.177163 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.177748 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.677724942 +0000 UTC m=+143.762262836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.178195 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.178537 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.678529063 +0000 UTC m=+143.763066947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.234809 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.247764 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.281851 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.281991 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.781963564 +0000 UTC m=+143.866501448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.282248 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.282691 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.782675323 +0000 UTC m=+143.867213207 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.288383 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.298039 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.330703 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.362382 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.375692 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bst8g" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.381344 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hr4hw" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.383341 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.383623 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.883608658 +0000 UTC m=+143.968146542 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.433815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" event={"ID":"2928d6d7-52d6-40ca-b351-3efbc1bfe6ff","Type":"ContainerStarted","Data":"eae32f04f2becd4138e7f1f6fa85dd42caae5188004b0bc7cedceba444922f8f"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.479000 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" event={"ID":"f36e10a7-e2e0-45d2-8ba1-a7fb742a3dbd","Type":"ContainerStarted","Data":"5953b3855690c081fc1990b41ed83d0425469943a7fff6d4d4702c25678b9703"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.484148 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.484453 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:55.984443051 +0000 UTC m=+144.068980935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.493879 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" event={"ID":"cf080121-ca40-4473-9ca6-a8feb0780bf2","Type":"ContainerStarted","Data":"8c0730fae0c16d6ee7b93a176e968b32ded5ebe2a8ca6af00b02aafcc96b70ec"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.497184 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" event={"ID":"6b9796cc-d9c6-445d-8dce-1c048c21603e","Type":"ContainerStarted","Data":"0b5049144062a503324e0d79b25ac2cbcfccb396b7a88b9e2094d4b736cffd24"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.501110 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c9rl2"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.506947 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.506981 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.529598 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.531783 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" event={"ID":"297a1de0-f405-4338-aeab-1bf11a7afc8c","Type":"ContainerStarted","Data":"d099c7f7e1e662e9c28ab84ad15257d6d3893b32fad2e45350c68fb237cf79b4"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.546003 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.548631 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.558413 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" event={"ID":"17cb88f3-b5b5-4d92-94c8-fdcdb873c224","Type":"ContainerStarted","Data":"dad0c6b983a5f28c162f36e0d24adc6175783abaf9233058f3d8e826343a198d"} Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.586033 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.586180 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.086154166 +0000 UTC m=+144.170692040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.586230 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.586555 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.086547927 +0000 UTC m=+144.171085811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.663728 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.695465 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.698093 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.19807413 +0000 UTC m=+144.282612014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.750792 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-8vwh8" podStartSLOduration=122.750775447 podStartE2EDuration="2m2.750775447s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:55.749998686 +0000 UTC m=+143.834536570" watchObservedRunningTime="2025-10-13 12:45:55.750775447 +0000 UTC m=+143.835313331" Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.767484 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj"] Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.797842 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.798740 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.298728678 +0000 UTC m=+144.383266562 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:55 crc kubenswrapper[4678]: I1013 12:45:55.899516 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:55 crc kubenswrapper[4678]: E1013 12:45:55.899863 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.399848789 +0000 UTC m=+144.484386673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.000887 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.001260 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.501247826 +0000 UTC m=+144.585785710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.035944 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" podStartSLOduration=123.035926238 podStartE2EDuration="2m3.035926238s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.008431035 +0000 UTC m=+144.092968919" watchObservedRunningTime="2025-10-13 12:45:56.035926238 +0000 UTC m=+144.120464112" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.101672 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.101986 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.601970915 +0000 UTC m=+144.686508799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: W1013 12:45:56.104721 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfb45e1a_27cb_4199_9a3d_657962b29d32.slice/crio-01333799daaba45dd3f234aea023684422ee5b17f250c1f198a98789489dbaa6 WatchSource:0}: Error finding container 01333799daaba45dd3f234aea023684422ee5b17f250c1f198a98789489dbaa6: Status 404 returned error can't find the container with id 01333799daaba45dd3f234aea023684422ee5b17f250c1f198a98789489dbaa6 Oct 13 12:45:56 crc kubenswrapper[4678]: W1013 12:45:56.166259 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod098d2671_5613_4906_af73_dc25b2ebd868.slice/crio-5d4c1cd576e7e5da247d56841b1fc061e0342f6bdd831ea03bb385d6ba6dc5d3 WatchSource:0}: Error finding container 5d4c1cd576e7e5da247d56841b1fc061e0342f6bdd831ea03bb385d6ba6dc5d3: Status 404 returned error can't find the container with id 5d4c1cd576e7e5da247d56841b1fc061e0342f6bdd831ea03bb385d6ba6dc5d3 Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.204401 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.204654 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.704644046 +0000 UTC m=+144.789181930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.311301 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.311762 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.811745494 +0000 UTC m=+144.896283378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.315430 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hkh7k"] Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.347423 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.412684 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.413325 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:56.913277175 +0000 UTC m=+144.997815059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.493741 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" podStartSLOduration=123.493705861 podStartE2EDuration="2m3.493705861s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.450392531 +0000 UTC m=+144.534930425" watchObservedRunningTime="2025-10-13 12:45:56.493705861 +0000 UTC m=+144.578243745" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.506114 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" podStartSLOduration=123.506086066 podStartE2EDuration="2m3.506086066s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.487326913 +0000 UTC m=+144.571864797" watchObservedRunningTime="2025-10-13 12:45:56.506086066 +0000 UTC m=+144.590623950" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.515319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.519104 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.019078758 +0000 UTC m=+145.103616642 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.552427 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" podStartSLOduration=123.552397244 podStartE2EDuration="2m3.552397244s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.536718432 +0000 UTC m=+144.621256316" watchObservedRunningTime="2025-10-13 12:45:56.552397244 +0000 UTC m=+144.636935118" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.611901 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" event={"ID":"098d2671-5613-4906-af73-dc25b2ebd868","Type":"ContainerStarted","Data":"5d4c1cd576e7e5da247d56841b1fc061e0342f6bdd831ea03bb385d6ba6dc5d3"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.618362 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.618816 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.118802871 +0000 UTC m=+145.203340755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.638387 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" event={"ID":"cf080121-ca40-4473-9ca6-a8feb0780bf2","Type":"ContainerStarted","Data":"3548958431c4d7b3252cb5d2e69543e72379922b44d1be429aaa381f7503074a"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.686997 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hr4hw" event={"ID":"916d3614-f8db-4446-bda3-224c9bc15357","Type":"ContainerStarted","Data":"211c636aff3535251d5267d1c67960621ab0dd5a706aaf596aeb2a02e8d49543"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.724421 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" event={"ID":"7b9d0828-76bc-451a-875a-bb2fb0f56322","Type":"ContainerStarted","Data":"c35f6cef288a2f1e71c1fa917021ca58eab6bbd68a02946226ef6e7f280d09fc"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.725303 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.726340 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.22632527 +0000 UTC m=+145.310863154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.758577 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-dkxs9" podStartSLOduration=123.758561598 podStartE2EDuration="2m3.758561598s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.757711845 +0000 UTC m=+144.842249729" watchObservedRunningTime="2025-10-13 12:45:56.758561598 +0000 UTC m=+144.843099482" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.760802 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" event={"ID":"297a1de0-f405-4338-aeab-1bf11a7afc8c","Type":"ContainerStarted","Data":"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.760874 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.809619 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq6wd" event={"ID":"bfb45e1a-27cb-4199-9a3d-657962b29d32","Type":"ContainerStarted","Data":"01333799daaba45dd3f234aea023684422ee5b17f250c1f198a98789489dbaa6"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.827577 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.831166 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.331147027 +0000 UTC m=+145.415684911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.835490 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" event={"ID":"c3d0f53a-d207-44be-a692-894d2740fa54","Type":"ContainerStarted","Data":"37f915a5e49896447c27f94f7472409a2d9b6857c3e686603bef9381293415c9"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.835527 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" event={"ID":"c3d0f53a-d207-44be-a692-894d2740fa54","Type":"ContainerStarted","Data":"051bffaac98ca65caf07cfc4b69513a2de99eb89ef7329a793ac63d4290a8a79"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.858551 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" event={"ID":"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a","Type":"ContainerStarted","Data":"8e6fdd57788beeaf612abe47dc1fd7aa75faf5a2c8fb2095672954e3829b047f"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.870519 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" event={"ID":"6b9796cc-d9c6-445d-8dce-1c048c21603e","Type":"ContainerStarted","Data":"fa5b12773958565e007387ac0611ba888b9275a0406b32cd95ac01546870a800"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.888173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tsjvf" event={"ID":"6e30bbf0-723d-4233-b05f-a679c0e32ad5","Type":"ContainerStarted","Data":"179cc20b8a524d48dc49349efad218077b9a906cf4c748cb5815fceb211179bf"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.888241 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tsjvf" event={"ID":"6e30bbf0-723d-4233-b05f-a679c0e32ad5","Type":"ContainerStarted","Data":"82f41ff6fda38e985360d586320ba839e8a60af8494f96b2b34854b7c2dc0e0f"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.903949 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hkh7k" event={"ID":"21bd5a7f-366f-4d05-986d-6285b95bb1a4","Type":"ContainerStarted","Data":"c847a7619f06da28e77930645668297b759ca41f1686b34aa3b98f54476937c1"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.905850 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" event={"ID":"5ea4f437-7feb-4ca7-9e73-631c8dad195e","Type":"ContainerStarted","Data":"8d655725dbfce6b2027f9574bdc38c2f816d9d35fc766e53a3e838186f56fb7a"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.912164 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" event={"ID":"17cb88f3-b5b5-4d92-94c8-fdcdb873c224","Type":"ContainerStarted","Data":"82f3b23edfd72633a1af626211795fd0ae11dc6d302897573afa845fe72ac852"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.919338 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c9rl2" event={"ID":"f35010e9-46ee-4ba5-8e85-b7568bf88650","Type":"ContainerStarted","Data":"2cb6107ec55d32a7f6975b33e0ecf0c6cf018b7934a0bdd4e60fb8eb9bf761ec"} Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.926833 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.929433 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:56 crc kubenswrapper[4678]: E1013 12:45:56.931021 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.430996184 +0000 UTC m=+145.515534068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.934728 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-q8tf2" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.947082 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-hzdvq" podStartSLOduration=124.947045546 podStartE2EDuration="2m4.947045546s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.93503363 +0000 UTC m=+145.019571534" watchObservedRunningTime="2025-10-13 12:45:56.947045546 +0000 UTC m=+145.031583430" Oct 13 12:45:56 crc kubenswrapper[4678]: I1013 12:45:56.979546 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" podStartSLOduration=124.97953058 podStartE2EDuration="2m4.97953058s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:56.971497929 +0000 UTC m=+145.056035813" watchObservedRunningTime="2025-10-13 12:45:56.97953058 +0000 UTC m=+145.064068464" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.030701 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.033745 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.533731085 +0000 UTC m=+145.618269069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.049287 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-tsjvf" podStartSLOduration=124.049268914 podStartE2EDuration="2m4.049268914s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.031733943 +0000 UTC m=+145.116271827" watchObservedRunningTime="2025-10-13 12:45:57.049268914 +0000 UTC m=+145.133806798" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.105011 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" podStartSLOduration=125.10498653 podStartE2EDuration="2m5.10498653s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.102798962 +0000 UTC m=+145.187336846" watchObservedRunningTime="2025-10-13 12:45:57.10498653 +0000 UTC m=+145.189524424" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.108642 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-v7vb4"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.133072 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.133659 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.633637013 +0000 UTC m=+145.718174887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.169616 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" podStartSLOduration=125.169593159 podStartE2EDuration="2m5.169593159s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.147651512 +0000 UTC m=+145.232189396" watchObservedRunningTime="2025-10-13 12:45:57.169593159 +0000 UTC m=+145.254131043" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.172256 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.235089 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.235379 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.236486 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.736468228 +0000 UTC m=+145.821006112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.238671 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-65hzw"] Oct 13 12:45:57 crc kubenswrapper[4678]: W1013 12:45:57.244834 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b555cbc_dd4d_45bc_b311_ac341169716b.slice/crio-0a6756d33e03b850ad2a9b4ce6a15e1ee364122c9f71ca482027ff6428fc0c19 WatchSource:0}: Error finding container 0a6756d33e03b850ad2a9b4ce6a15e1ee364122c9f71ca482027ff6428fc0c19: Status 404 returned error can't find the container with id 0a6756d33e03b850ad2a9b4ce6a15e1ee364122c9f71ca482027ff6428fc0c19 Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.248783 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.261601 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.261651 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.266777 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm2j9"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.277648 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv"] Oct 13 12:45:57 crc kubenswrapper[4678]: W1013 12:45:57.291503 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48f355d2_c241_44aa_a567_e9c31de5d18e.slice/crio-2fcc6df91b8b2de75037279201ec6760265c418a689fa481dea7ac7906de5f01 WatchSource:0}: Error finding container 2fcc6df91b8b2de75037279201ec6760265c418a689fa481dea7ac7906de5f01: Status 404 returned error can't find the container with id 2fcc6df91b8b2de75037279201ec6760265c418a689fa481dea7ac7906de5f01 Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.298127 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-c9rl2" podStartSLOduration=124.29810584 podStartE2EDuration="2m4.29810584s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.181504322 +0000 UTC m=+145.266042206" watchObservedRunningTime="2025-10-13 12:45:57.29810584 +0000 UTC m=+145.382643724" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.321986 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.340217 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.340848 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.341169 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.841152452 +0000 UTC m=+145.925690336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.350499 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qnj4p"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.359414 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" podStartSLOduration=124.359390012 podStartE2EDuration="2m4.359390012s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.241166172 +0000 UTC m=+145.325704046" watchObservedRunningTime="2025-10-13 12:45:57.359390012 +0000 UTC m=+145.443927896" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.362442 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-qq6wd" podStartSLOduration=124.362432742 podStartE2EDuration="2m4.362432742s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.279043378 +0000 UTC m=+145.363581262" watchObservedRunningTime="2025-10-13 12:45:57.362432742 +0000 UTC m=+145.446970626" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.388159 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5hccp"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.391514 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ldbrg" podStartSLOduration=124.391492197 podStartE2EDuration="2m4.391492197s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.350456757 +0000 UTC m=+145.434994641" watchObservedRunningTime="2025-10-13 12:45:57.391492197 +0000 UTC m=+145.476030081" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.392913 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.398824 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wmbl2" podStartSLOduration=124.398804859 podStartE2EDuration="2m4.398804859s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:57.373434592 +0000 UTC m=+145.457972486" watchObservedRunningTime="2025-10-13 12:45:57.398804859 +0000 UTC m=+145.483342743" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.425278 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.451853 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.452323 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:57.952309087 +0000 UTC m=+146.036846971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.474670 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.496090 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.511402 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.522771 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vjqg7"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.568016 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.568512 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.068492243 +0000 UTC m=+146.153030127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.571690 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-nggsb"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.621892 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bst8g"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.631109 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.639538 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp"] Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.673849 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.676306 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.176290849 +0000 UTC m=+146.260828733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: W1013 12:45:57.712837 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02bee92f_4cf8_476d_8d83_8eddce9a34ab.slice/crio-6332f26091452b07de314313e4de01be312b64224b54ba8d2f81df59eef5f57e WatchSource:0}: Error finding container 6332f26091452b07de314313e4de01be312b64224b54ba8d2f81df59eef5f57e: Status 404 returned error can't find the container with id 6332f26091452b07de314313e4de01be312b64224b54ba8d2f81df59eef5f57e Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.774709 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.775152 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.275136599 +0000 UTC m=+146.359674483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.797810 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.798081 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.880583 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.881309 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.381296491 +0000 UTC m=+146.465834375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.903211 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.903411 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.918560 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.944129 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" event={"ID":"80adeb99-a848-4261-a789-70cd9c333430","Type":"ContainerStarted","Data":"513af4ea0d66482d8f46f985bcf697c8b38acb18dd28483ad1afcd78d52ae397"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.946302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" event={"ID":"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b","Type":"ContainerStarted","Data":"ba4cb436217a48a0a977f3d2e92ab1781c5154043f9609de9b0f590b84dfc389"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.954135 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" event={"ID":"48f355d2-c241-44aa-a567-e9c31de5d18e","Type":"ContainerStarted","Data":"2fcc6df91b8b2de75037279201ec6760265c418a689fa481dea7ac7906de5f01"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.965470 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hkh7k" event={"ID":"21bd5a7f-366f-4d05-986d-6285b95bb1a4","Type":"ContainerStarted","Data":"8a760d07f8314d4fca26972be509b56f52cafb098b5879d12208ab7897c727d3"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.965519 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hkh7k" event={"ID":"21bd5a7f-366f-4d05-986d-6285b95bb1a4","Type":"ContainerStarted","Data":"47d7d28bcd7d852eed4534bfd23e4335fb63e1ff91cd15c7f1a3c0166e8c2f96"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.966084 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hkh7k" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.967987 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" event={"ID":"e8b410b3-0ca3-419a-9dcb-188b954a2136","Type":"ContainerStarted","Data":"5e1491aef4db57258ccb835e740ea456b32320dc38abd23167b9a3c1cd891c64"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.969740 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" event={"ID":"ad4e211d-f191-4262-9e70-d1f4fcaddeff","Type":"ContainerStarted","Data":"99a7417bd188db262347d6ad5de564f7d67778c5d172d3fa65035468d25d381b"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.970407 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" event={"ID":"56d1be59-35a1-44ac-873a-d5bc3df2f9ab","Type":"ContainerStarted","Data":"0b3040f61be7018f4048788b78efdba644ccca8bd0aa8088ac26ca9dc0b0697b"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.971381 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" event={"ID":"1ba5a0dd-a844-42cf-80f0-1c49802e73e0","Type":"ContainerStarted","Data":"bc75c2a9d394143eaf7775c5b276ab114ed23c5b9317b139fb7502c8b3c78102"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.972350 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" event={"ID":"098d2671-5613-4906-af73-dc25b2ebd868","Type":"ContainerStarted","Data":"26b7131060608e7650b4a28f809f5befaccb708b0f92f20512046faec3c05d44"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.973045 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.975744 4678 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xlrtj container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.975828 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" podUID="098d2671-5613-4906-af73-dc25b2ebd868" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.976733 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" event={"ID":"5632c50a-0b60-417c-afc1-633d42baf685","Type":"ContainerStarted","Data":"75b2133b36641a799a227c050e12771d1e9ec0b8e3162350348f1767aad153f8"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.976764 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" event={"ID":"5632c50a-0b60-417c-afc1-633d42baf685","Type":"ContainerStarted","Data":"6aabf66742af83b1066ebdd8b7d9c76f3291261a476fd394d1fab701436416c5"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.978098 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hr4hw" event={"ID":"916d3614-f8db-4446-bda3-224c9bc15357","Type":"ContainerStarted","Data":"6f62d167682afb3d788d11de08802e06d8bf9f875c3c8dc6bdface5829efbcd6"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.982262 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:57 crc kubenswrapper[4678]: E1013 12:45:57.982563 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.482548375 +0000 UTC m=+146.567086259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.982615 4678 patch_prober.go:28] interesting pod/apiserver-76f77b778f-shbr8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]log ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]etcd ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/generic-apiserver-start-informers ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/max-in-flight-filter ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 13 12:45:57 crc kubenswrapper[4678]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 13 12:45:57 crc kubenswrapper[4678]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/project.openshift.io-projectcache ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/openshift.io-startinformers ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 13 12:45:57 crc kubenswrapper[4678]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 13 12:45:57 crc kubenswrapper[4678]: livez check failed Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.982678 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" podUID="17cb88f3-b5b5-4d92-94c8-fdcdb873c224" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.993879 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" event={"ID":"7b9d0828-76bc-451a-875a-bb2fb0f56322","Type":"ContainerStarted","Data":"21bf8043b79670ea121aae3cdc6150969524af9a72e5fa950e3d475be2de5346"} Oct 13 12:45:57 crc kubenswrapper[4678]: I1013 12:45:57.993939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" event={"ID":"7b9d0828-76bc-451a-875a-bb2fb0f56322","Type":"ContainerStarted","Data":"1514c454d122f251776cce24840b85ba4c5c04cfb5e37f63accfeb623d235f2a"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.021355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" event={"ID":"99ba1b17-157b-4df8-b753-52b5742443cd","Type":"ContainerStarted","Data":"a59c587f5a978bd2266e83734878ca498662c5806d94d4f9660cbc9ad1331bde"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.021684 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" event={"ID":"99ba1b17-157b-4df8-b753-52b5742443cd","Type":"ContainerStarted","Data":"ae2feb7c1617187977b8cce393f9346559c8ae0ad76af4aa3c06ab4c27ae9338"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.023365 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.023426 4678 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-9j7x8 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.023452 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" podUID="99ba1b17-157b-4df8-b753-52b5742443cd" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.038256 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" event={"ID":"96d06121-ad51-4e70-9a3a-952c6863efed","Type":"ContainerStarted","Data":"1eac5a90ff1acfd1ecccbe5ae06a93f24194332fe69304ba81683c55ec68ad06"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.038302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" event={"ID":"96d06121-ad51-4e70-9a3a-952c6863efed","Type":"ContainerStarted","Data":"c58b28adbc20c1194d94fc16c2b81f8f616222eee4a2cfc537278bfb5f415681"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.043967 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hkh7k" podStartSLOduration=7.04395541 podStartE2EDuration="7.04395541s" podCreationTimestamp="2025-10-13 12:45:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.004728878 +0000 UTC m=+146.089266762" watchObservedRunningTime="2025-10-13 12:45:58.04395541 +0000 UTC m=+146.128493284" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.048043 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" event={"ID":"5ea4f437-7feb-4ca7-9e73-631c8dad195e","Type":"ContainerStarted","Data":"d37075c41d89347fe1278458a6c13a546f4b2da41a7b17924979d3b939979bf3"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.048099 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" event={"ID":"5ea4f437-7feb-4ca7-9e73-631c8dad195e","Type":"ContainerStarted","Data":"ff5d083157fccd9f2d140aa81174a85a0b0a19b9a08ad37e0f70646006c003fd"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.054609 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" event={"ID":"839c0ab6-3764-4934-9d39-a09227958854","Type":"ContainerStarted","Data":"fcac5aac0b5d0af0677d288334e27317db41e45b4d6c2620062e9ba4c7160356"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.057437 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" event={"ID":"dbf24aeb-11f0-4642-ac35-c69bd78242b5","Type":"ContainerStarted","Data":"180c32f25d13b7b985c26c3d9a20bc796b2863ef859a9707a518df5f4366b210"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.063360 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" event={"ID":"a497df52-2462-4299-b266-6f5ef91edaef","Type":"ContainerStarted","Data":"682a676ec9f1e94e6202fa4b7c22674f67f5de84af3d724788c970e91e3f87d0"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.063395 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" event={"ID":"a497df52-2462-4299-b266-6f5ef91edaef","Type":"ContainerStarted","Data":"959675df2e570560a51a6546009abe6f04b6267dc043a9eb7b9547012ce73f3e"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.075915 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hr4hw" podStartSLOduration=6.075896581 podStartE2EDuration="6.075896581s" podCreationTimestamp="2025-10-13 12:45:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.075356356 +0000 UTC m=+146.159894230" watchObservedRunningTime="2025-10-13 12:45:58.075896581 +0000 UTC m=+146.160434465" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.076761 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" podStartSLOduration=126.076752983 podStartE2EDuration="2m6.076752983s" podCreationTimestamp="2025-10-13 12:43:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.047609316 +0000 UTC m=+146.132147200" watchObservedRunningTime="2025-10-13 12:45:58.076752983 +0000 UTC m=+146.161290867" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.085082 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" event={"ID":"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1","Type":"ContainerStarted","Data":"46b9e47b9f978c5cb1e6a6368f0dec10bdb6e8ca1d0d3d99542b2fc661963eb3"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.086543 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.092658 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.592642671 +0000 UTC m=+146.677180545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.115172 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" event={"ID":"6b555cbc-dd4d-45bc-b311-ac341169716b","Type":"ContainerStarted","Data":"0a6756d33e03b850ad2a9b4ce6a15e1ee364122c9f71ca482027ff6428fc0c19"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.116711 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.136233 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" podStartSLOduration=125.136215487 podStartE2EDuration="2m5.136215487s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.134554324 +0000 UTC m=+146.219092208" watchObservedRunningTime="2025-10-13 12:45:58.136215487 +0000 UTC m=+146.220753361" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.138366 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-w87mp" event={"ID":"8d82b4dc-c0dc-4672-bd1f-b7a2be4fbd9a","Type":"ContainerStarted","Data":"e06c903106b79d686c2ab5d92a4840bccf6493c53031c3dbd0806e28b66501fb"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.168884 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pj7h9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.168941 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.185411 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8hr85" podStartSLOduration=125.185395911 podStartE2EDuration="2m5.185395911s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.183921312 +0000 UTC m=+146.268459196" watchObservedRunningTime="2025-10-13 12:45:58.185395911 +0000 UTC m=+146.269933795" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.187982 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.190282 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.690262179 +0000 UTC m=+146.774800063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.191831 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bst8g" event={"ID":"02bee92f-4cf8-476d-8d83-8eddce9a34ab","Type":"ContainerStarted","Data":"6332f26091452b07de314313e4de01be312b64224b54ba8d2f81df59eef5f57e"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.252747 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:45:58 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:45:58 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:45:58 crc kubenswrapper[4678]: healthz check failed Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.252802 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.257853 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-65hzw" podStartSLOduration=125.257837877 podStartE2EDuration="2m5.257837877s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.237029979 +0000 UTC m=+146.321567863" watchObservedRunningTime="2025-10-13 12:45:58.257837877 +0000 UTC m=+146.342375761" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.267813 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq6wd" event={"ID":"bfb45e1a-27cb-4199-9a3d-657962b29d32","Type":"ContainerStarted","Data":"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.293810 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-v7vb4" podStartSLOduration=125.293781952 podStartE2EDuration="2m5.293781952s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.257361194 +0000 UTC m=+146.341899078" watchObservedRunningTime="2025-10-13 12:45:58.293781952 +0000 UTC m=+146.378319836" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.294383 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.294682 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.794670646 +0000 UTC m=+146.879208520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.296165 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rqc8w" podStartSLOduration=125.296158015 podStartE2EDuration="2m5.296158015s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.294032059 +0000 UTC m=+146.378569943" watchObservedRunningTime="2025-10-13 12:45:58.296158015 +0000 UTC m=+146.380695899" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.307136 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" event={"ID":"80104b9e-1a05-40d6-b179-f9330679357a","Type":"ContainerStarted","Data":"8c6d53281ce63cb90ece724da784cefaf866ecac112f19768b99fe4d6ede5e31"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.322735 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c9rl2" event={"ID":"f35010e9-46ee-4ba5-8e85-b7568bf88650","Type":"ContainerStarted","Data":"09b16c290f71be380dd1aa6f6b23c07ba0d53c4825849f747c905e71c41e2929"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.323773 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.327508 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-b5mc9" podStartSLOduration=125.327497899 podStartE2EDuration="2m5.327497899s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.324512551 +0000 UTC m=+146.409050445" watchObservedRunningTime="2025-10-13 12:45:58.327497899 +0000 UTC m=+146.412035783" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.339721 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-c9rl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.340135 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c9rl2" podUID="f35010e9-46ee-4ba5-8e85-b7568bf88650" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.340447 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" event={"ID":"d6dc2394-2372-4e43-ae9f-8557a979464c","Type":"ContainerStarted","Data":"cb2fc5c17567a93245fb7b6ef8bb516fd19dbd52457ba1fd8bbf728fbaa345ab"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.353437 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" event={"ID":"b79db0ac-d465-4200-acc8-194524524a3a","Type":"ContainerStarted","Data":"eb1283c141edd2b64b36dba6e1f22558670b3922cb129f74034f20a91c63aa75"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.357312 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" event={"ID":"214c4966-ab84-4ea0-acf4-8446bd0f2cd0","Type":"ContainerStarted","Data":"bf4806c53574d525900d027d0e00b71ac8e457a067dc8724a9dbc8ad672a4a2d"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.362875 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" podStartSLOduration=125.362859369 podStartE2EDuration="2m5.362859369s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.361934745 +0000 UTC m=+146.446472629" watchObservedRunningTime="2025-10-13 12:45:58.362859369 +0000 UTC m=+146.447397243" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.367683 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cf28k" event={"ID":"6b9796cc-d9c6-445d-8dce-1c048c21603e","Type":"ContainerStarted","Data":"ad8901dff185ebc77ba2d5abc906060e9d1687ef6dc4eb23458760239fd74a71"} Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.380134 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gr9dd" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.384069 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" podStartSLOduration=125.384029596 podStartE2EDuration="2m5.384029596s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:58.383239546 +0000 UTC m=+146.467777430" watchObservedRunningTime="2025-10-13 12:45:58.384029596 +0000 UTC m=+146.468567480" Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.395536 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.395691 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.895665173 +0000 UTC m=+146.980203057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.395961 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.397606 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.897595653 +0000 UTC m=+146.982133537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.497175 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.498229 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:58.99820272 +0000 UTC m=+147.082740594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.603423 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.604347 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.104331532 +0000 UTC m=+147.188869416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.705508 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.705890 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.205848793 +0000 UTC m=+147.290386677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.807904 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.808358 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.308347579 +0000 UTC m=+147.392885463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.909013 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.909291 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.409264964 +0000 UTC m=+147.493802848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:58 crc kubenswrapper[4678]: I1013 12:45:58.909537 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:58 crc kubenswrapper[4678]: E1013 12:45:58.910120 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.410106056 +0000 UTC m=+147.494643940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.011150 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.011341 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.511311389 +0000 UTC m=+147.595849273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.011557 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.011930 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.511921105 +0000 UTC m=+147.596458989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.112563 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.112701 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.612673835 +0000 UTC m=+147.697211719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.112850 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.113216 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.613207649 +0000 UTC m=+147.697745533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.214063 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.214232 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.714197536 +0000 UTC m=+147.798735420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.214652 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.215021 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.715011637 +0000 UTC m=+147.799549521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.255404 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:45:59 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:45:59 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:45:59 crc kubenswrapper[4678]: healthz check failed Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.255912 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.316376 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.316700 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.816671912 +0000 UTC m=+147.901209796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.317243 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.317614 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.817606086 +0000 UTC m=+147.902143970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.374254 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" event={"ID":"b79db0ac-d465-4200-acc8-194524524a3a","Type":"ContainerStarted","Data":"90108e0c715936aaef5b1f49977772704d6d176db7e8ef6382e4198b1ef08ce6"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.374868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" event={"ID":"b79db0ac-d465-4200-acc8-194524524a3a","Type":"ContainerStarted","Data":"5ffb52c5149ff6d155a755f9227f9572ae6ba0ea45760f76f8b88c2df8464fda"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.376520 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" event={"ID":"214c4966-ab84-4ea0-acf4-8446bd0f2cd0","Type":"ContainerStarted","Data":"01b18aae6892116a658860cd74f65cda716301d81811deaf6469843f3edb3567"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.376570 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" event={"ID":"214c4966-ab84-4ea0-acf4-8446bd0f2cd0","Type":"ContainerStarted","Data":"8527a1d1a547242df77061477d2b01f9623d89dc1a4f5fd217c49d119a431235"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.380854 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.381478 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.382113 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" event={"ID":"839c0ab6-3764-4934-9d39-a09227958854","Type":"ContainerStarted","Data":"599b14dff31b708d53dfdc604db79058284734caa5750a2edfc47cba138d8985"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.383377 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.383538 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.384164 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" event={"ID":"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b","Type":"ContainerStarted","Data":"c541b4b595ecbe957a63bf79ee6b038816fb0db57304bff1a17464b372a01536"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.384198 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" event={"ID":"65ee6db4-4e7e-41e8-9a6f-14c54e13f36b","Type":"ContainerStarted","Data":"81b082d33a52774063db59ca01b8919589f0accc9e2aaaae5322572ca6f9ddac"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.388444 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.392002 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bst8g" event={"ID":"02bee92f-4cf8-476d-8d83-8eddce9a34ab","Type":"ContainerStarted","Data":"9e30b5cc1f78780b6a97bcb939d1d608ea5ac7e20b12a753626f927e480d1237"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.394386 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" event={"ID":"d6dc2394-2372-4e43-ae9f-8557a979464c","Type":"ContainerStarted","Data":"b4f48e1f76f46ab74ea9d8d03576290d9408c8a5cf1423ea7c255fe232a91ba8"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.399790 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" event={"ID":"1ba5a0dd-a844-42cf-80f0-1c49802e73e0","Type":"ContainerStarted","Data":"7ecaa9c13fde2e7dbb3379c4f08db9d18bd6adb5ef3b4554054472fcd6af9930"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.399844 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" event={"ID":"1ba5a0dd-a844-42cf-80f0-1c49802e73e0","Type":"ContainerStarted","Data":"53dad553635890826f6a6ae16232b2021631ce7e65f82fd97a60eaa7ed82232b"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.399886 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.401564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" event={"ID":"80104b9e-1a05-40d6-b179-f9330679357a","Type":"ContainerStarted","Data":"c8db766319bd7551db32eaf0f38812d14859600cc2736294fb833e19df2a8a6f"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.402772 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" event={"ID":"3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1","Type":"ContainerStarted","Data":"8ba7f2d28cf9221279fd9b21990d14c057a40a8f928b60a5fe1795bddf0078ff"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.402881 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.404428 4678 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-xbd8l container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.404549 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" podUID="3a18c90b-ce8b-4cea-98f5-31bc0e25d3b1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.404487 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5hccp" event={"ID":"e8b410b3-0ca3-419a-9dcb-188b954a2136","Type":"ContainerStarted","Data":"ba3b897838fc313ac11efaaef5d0ae88f0b05acb7630ccf792534cf7f1833615"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.405845 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" event={"ID":"56d1be59-35a1-44ac-873a-d5bc3df2f9ab","Type":"ContainerStarted","Data":"0cc9073e2d3cf3fa791a0550e9009bb166d58a4a15b452a8d15d89794b6b0c39"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.406907 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" event={"ID":"80adeb99-a848-4261-a789-70cd9c333430","Type":"ContainerStarted","Data":"58e8df3c860587c8528e7ac633eb244c92c802da0477dc440dc28f322ea03404"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.408891 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" event={"ID":"dbf24aeb-11f0-4642-ac35-c69bd78242b5","Type":"ContainerStarted","Data":"d3a5f189cfa070a48a469db4648c0b1c5af65ff8ecd7f5df65146b2645b201fe"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.408956 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" event={"ID":"dbf24aeb-11f0-4642-ac35-c69bd78242b5","Type":"ContainerStarted","Data":"e234c8d1f9e87a584be61ca1428232994f983d84dd01bbfbae279bdf75481366"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.410045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" event={"ID":"6b555cbc-dd4d-45bc-b311-ac341169716b","Type":"ContainerStarted","Data":"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.410723 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-pj7h9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.410765 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.412657 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" event={"ID":"48f355d2-c241-44aa-a567-e9c31de5d18e","Type":"ContainerStarted","Data":"72bf7625f8d2f4f271625881011d0d7bdf63d0b48db770c6c0f35933140e329f"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.412934 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.414393 4678 patch_prober.go:28] interesting pod/console-operator-58897d9998-qnj4p container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.414423 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" podUID="48f355d2-c241-44aa-a567-e9c31de5d18e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.414508 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" event={"ID":"ad4e211d-f191-4262-9e70-d1f4fcaddeff","Type":"ContainerStarted","Data":"b7b0444363b948c48374c9e2716907db54d911082b0c90ea7c4478d80d6b0bcd"} Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.415183 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-c9rl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.415231 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c9rl2" podUID="f35010e9-46ee-4ba5-8e85-b7568bf88650" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.427450 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.428932 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:45:59.928914084 +0000 UTC m=+148.013451968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.440616 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xlrtj" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.453357 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8cp6h" podStartSLOduration=126.453325847 podStartE2EDuration="2m6.453325847s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.430045134 +0000 UTC m=+147.514583028" watchObservedRunningTime="2025-10-13 12:45:59.453325847 +0000 UTC m=+147.537863731" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.501015 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" podStartSLOduration=59.50098854 podStartE2EDuration="59.50098854s" podCreationTimestamp="2025-10-13 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.4770206 +0000 UTC m=+147.561558474" watchObservedRunningTime="2025-10-13 12:45:59.50098854 +0000 UTC m=+147.585526414" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.515482 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bgtc" podStartSLOduration=126.515470751 podStartE2EDuration="2m6.515470751s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.500677462 +0000 UTC m=+147.585215346" watchObservedRunningTime="2025-10-13 12:45:59.515470751 +0000 UTC m=+147.600008635" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.532327 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.532710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.532762 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.539850 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jbw5p" podStartSLOduration=126.539822982 podStartE2EDuration="2m6.539822982s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.527671382 +0000 UTC m=+147.612209266" watchObservedRunningTime="2025-10-13 12:45:59.539822982 +0000 UTC m=+147.624360866" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.544668 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.044651739 +0000 UTC m=+148.129189623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.568405 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" podStartSLOduration=126.568360943 podStartE2EDuration="2m6.568360943s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.559425118 +0000 UTC m=+147.643963002" watchObservedRunningTime="2025-10-13 12:45:59.568360943 +0000 UTC m=+147.652898837" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.620965 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jm2j9" podStartSLOduration=126.620850054 podStartE2EDuration="2m6.620850054s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.592698363 +0000 UTC m=+147.677236247" watchObservedRunningTime="2025-10-13 12:45:59.620850054 +0000 UTC m=+147.705387958" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.634080 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.634259 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.134227215 +0000 UTC m=+148.218765099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.634750 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.634862 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.634939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.635345 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.635716 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.135701664 +0000 UTC m=+148.220239548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.653488 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" podStartSLOduration=126.653469012 podStartE2EDuration="2m6.653469012s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.628042633 +0000 UTC m=+147.712580527" watchObservedRunningTime="2025-10-13 12:45:59.653469012 +0000 UTC m=+147.738006896" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.653882 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-nggsb" podStartSLOduration=126.653878402 podStartE2EDuration="2m6.653878402s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.651645374 +0000 UTC m=+147.736183258" watchObservedRunningTime="2025-10-13 12:45:59.653878402 +0000 UTC m=+147.738416286" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.673896 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.700434 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-6dcl8" podStartSLOduration=126.700418697 podStartE2EDuration="2m6.700418697s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.681505069 +0000 UTC m=+147.766042953" watchObservedRunningTime="2025-10-13 12:45:59.700418697 +0000 UTC m=+147.784956581" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.701012 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.701877 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.702250 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.705999 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.715120 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-m99jv" podStartSLOduration=126.715104813 podStartE2EDuration="2m6.715104813s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.714744994 +0000 UTC m=+147.799282878" watchObservedRunningTime="2025-10-13 12:45:59.715104813 +0000 UTC m=+147.799642697" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.721372 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.743567 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.743795 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nspzk\" (UniqueName: \"kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.743841 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.743883 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.744277 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.244220709 +0000 UTC m=+148.328758593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.755534 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bst8g" podStartSLOduration=8.755513266 podStartE2EDuration="8.755513266s" podCreationTimestamp="2025-10-13 12:45:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.744181198 +0000 UTC m=+147.828719082" watchObservedRunningTime="2025-10-13 12:45:59.755513266 +0000 UTC m=+147.840051150" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.771452 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cm2qp" podStartSLOduration=126.771433245 podStartE2EDuration="2m6.771433245s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.77012197 +0000 UTC m=+147.854659854" watchObservedRunningTime="2025-10-13 12:45:59.771433245 +0000 UTC m=+147.855971129" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845410 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845472 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845492 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845543 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845559 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nspzk\" (UniqueName: \"kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845579 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845601 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.845620 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.846419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.846672 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.346660954 +0000 UTC m=+148.431198838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.846979 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.848344 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.853635 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.868870 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" podStartSLOduration=126.868853628 podStartE2EDuration="2m6.868853628s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:45:59.863399584 +0000 UTC m=+147.947937458" watchObservedRunningTime="2025-10-13 12:45:59.868853628 +0000 UTC m=+147.953391512" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.870117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.871714 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.876753 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nspzk\" (UniqueName: \"kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk\") pod \"certified-operators-7jh7p\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.913016 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.913935 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.920533 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.955124 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.957959 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.958222 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.958249 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfng\" (UniqueName: \"kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:45:59 crc kubenswrapper[4678]: I1013 12:45:59.958490 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:45:59 crc kubenswrapper[4678]: E1013 12:45:59.958594 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.458577258 +0000 UTC m=+148.543115142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.058778 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.061260 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.061306 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.061325 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfng\" (UniqueName: \"kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.061343 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.061671 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.5616595 +0000 UTC m=+148.646197384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.062231 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.062515 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.080097 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfng\" (UniqueName: \"kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng\") pod \"community-operators-pcx8f\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.090283 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.113752 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.115483 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.121959 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.132365 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-9j7x8" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.133110 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.153739 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.162833 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.163166 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.163227 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.163258 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkxlq\" (UniqueName: \"kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.163418 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.663401886 +0000 UTC m=+148.747939770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.240771 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.257266 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:46:00 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:46:00 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:46:00 crc kubenswrapper[4678]: healthz check failed Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.257328 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.264814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.264860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.264884 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.264910 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkxlq\" (UniqueName: \"kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.265529 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.265753 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.765742869 +0000 UTC m=+148.850280743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.265953 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.314649 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.316724 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.321399 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkxlq\" (UniqueName: \"kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq\") pod \"certified-operators-p5f6j\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.339124 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.365842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.366079 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.366129 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtl79\" (UniqueName: \"kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.366201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.366300 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.866286754 +0000 UTC m=+148.950824628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.436452 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.444870 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-c9rl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.444920 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c9rl2" podUID="f35010e9-46ee-4ba5-8e85-b7568bf88650" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.458951 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.465419 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-xbd8l" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.467330 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.467385 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.467478 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.467545 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtl79\" (UniqueName: \"kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.471429 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.473998 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:00.973983977 +0000 UTC m=+149.058521861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.478484 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.503827 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qnj4p" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.524334 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtl79\" (UniqueName: \"kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79\") pod \"community-operators-2rk4n\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.548632 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.568847 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.569395 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.069369735 +0000 UTC m=+149.153907619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.653520 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.670698 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.671942 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.171924403 +0000 UTC m=+149.256462287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.772686 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.772982 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.272964611 +0000 UTC m=+149.357502495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.830722 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.885556 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.885975 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.385959953 +0000 UTC m=+149.470497837 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:00 crc kubenswrapper[4678]: I1013 12:46:00.986703 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:00 crc kubenswrapper[4678]: E1013 12:46:00.987288 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.487273589 +0000 UTC m=+149.571811473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.089419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.090030 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.589803676 +0000 UTC m=+149.674341560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.162202 4678 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.191998 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.192383 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.692363894 +0000 UTC m=+149.776901778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.194569 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:46:01 crc kubenswrapper[4678]: W1013 12:46:01.215680 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode30f6aef_1479_4509_9515_f65af72e1c33.slice/crio-c6c5409b0ac8be52cdc99978807ea5dcdfae9ee0cf833f8bf1c1dbafcbf2aead WatchSource:0}: Error finding container c6c5409b0ac8be52cdc99978807ea5dcdfae9ee0cf833f8bf1c1dbafcbf2aead: Status 404 returned error can't find the container with id c6c5409b0ac8be52cdc99978807ea5dcdfae9ee0cf833f8bf1c1dbafcbf2aead Oct 13 12:46:01 crc kubenswrapper[4678]: W1013 12:46:01.219404 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b258b71ec193d01d6a337957f4e3339f84351be561bae585686456508689cbc6 WatchSource:0}: Error finding container b258b71ec193d01d6a337957f4e3339f84351be561bae585686456508689cbc6: Status 404 returned error can't find the container with id b258b71ec193d01d6a337957f4e3339f84351be561bae585686456508689cbc6 Oct 13 12:46:01 crc kubenswrapper[4678]: W1013 12:46:01.240487 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-49bc3e0620789c2876bdd7b8df51ba55ab0e88282e92a7537698c3e1d91e1c8b WatchSource:0}: Error finding container 49bc3e0620789c2876bdd7b8df51ba55ab0e88282e92a7537698c3e1d91e1c8b: Status 404 returned error can't find the container with id 49bc3e0620789c2876bdd7b8df51ba55ab0e88282e92a7537698c3e1d91e1c8b Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.252072 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:46:01 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:46:01 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:46:01 crc kubenswrapper[4678]: healthz check failed Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.252376 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.297420 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.298935 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.798901617 +0000 UTC m=+149.883439501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.320525 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.334364 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:01 crc kubenswrapper[4678]: W1013 12:46:01.346517 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd2343db_8449_4042_bf5e_b1441b226020.slice/crio-6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390 WatchSource:0}: Error finding container 6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390: Status 404 returned error can't find the container with id 6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390 Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.399816 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.400230 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:01.900214162 +0000 UTC m=+149.984752046 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.460234 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerStarted","Data":"c6c5409b0ac8be52cdc99978807ea5dcdfae9ee0cf833f8bf1c1dbafcbf2aead"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.463253 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerID="cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e" exitCode=0 Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.463317 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerDied","Data":"cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.463336 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerStarted","Data":"7edd5476c30dc3cc1212ae750d2b71acd6d8a42a795ec9d06d603cdb1e3aad7a"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.465336 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.469999 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"49bc3e0620789c2876bdd7b8df51ba55ab0e88282e92a7537698c3e1d91e1c8b"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.475442 4678 generic.go:334] "Generic (PLEG): container finished" podID="839c0ab6-3764-4934-9d39-a09227958854" containerID="599b14dff31b708d53dfdc604db79058284734caa5750a2edfc47cba138d8985" exitCode=0 Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.475499 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" event={"ID":"839c0ab6-3764-4934-9d39-a09227958854","Type":"ContainerDied","Data":"599b14dff31b708d53dfdc604db79058284734caa5750a2edfc47cba138d8985"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.488465 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7","Type":"ContainerStarted","Data":"14b85184f2bdeb4ab78a1e557551ba79024f2ecfd49b7bfedc0b91cf79029e95"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.488501 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7","Type":"ContainerStarted","Data":"94800d6b24d17822281ade95dfe6152e26d90f988e37759a6214454b9d0e1dc8"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.491501 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerStarted","Data":"8159176dfeaf88a7365fd734a658927d211348e9a245f188786472bc3aad9ae2"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.495214 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" event={"ID":"56d1be59-35a1-44ac-873a-d5bc3df2f9ab","Type":"ContainerStarted","Data":"d2d4a6cb7e0c5ede84ebaf6cf50e035842657d939830d72b4572594770b7c344"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.495250 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" event={"ID":"56d1be59-35a1-44ac-873a-d5bc3df2f9ab","Type":"ContainerStarted","Data":"a922760d193223ed648ddfe9285bd4c23d28dbc14e6220864961f1dad48a677d"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.496837 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerStarted","Data":"6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.498413 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"226db8a17182057960d67051ce77aec80be858806c03de0756d553901422d4df"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.498443 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d75cdd5b9ea05efcd365aec6e701bf919ab9289fae759492f688de6c31fc5700"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.501304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.502118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ba6dfa236d91cef7ab10d260b6994761e4ae3754d160f87812ab0809a1bd5a28"} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.502154 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b258b71ec193d01d6a337957f4e3339f84351be561bae585686456508689cbc6"} Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.502587 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:02.002572314 +0000 UTC m=+150.087110198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.539668 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.53965151 podStartE2EDuration="2.53965151s" podCreationTimestamp="2025-10-13 12:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:46:01.523539366 +0000 UTC m=+149.608077250" watchObservedRunningTime="2025-10-13 12:46:01.53965151 +0000 UTC m=+149.624189394" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.602885 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.603079 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:02.103043507 +0000 UTC m=+150.187581391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.603479 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.604360 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:02.104353042 +0000 UTC m=+150.188890926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.705258 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.705438 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-13 12:46:02.20540169 +0000 UTC m=+150.289939614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.705500 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: E1013 12:46:01.706235 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-13 12:46:02.206220642 +0000 UTC m=+150.290758566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-929f2" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.768162 4678 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-13T12:46:01.162226531Z","Handler":null,"Name":""} Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.771773 4678 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.771817 4678 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.806973 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.810494 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.895641 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.902304 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.904457 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.904543 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.909455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.925033 4678 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.925087 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:01 crc kubenswrapper[4678]: I1013 12:46:01.969738 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-929f2\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.011032 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.011090 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xw9f\" (UniqueName: \"kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.011118 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.079267 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.114241 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.112952 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.115268 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xw9f\" (UniqueName: \"kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.115800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.116509 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.144147 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xw9f\" (UniqueName: \"kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f\") pod \"redhat-marketplace-9vhbd\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.230826 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.260823 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:46:02 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:46:02 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:46:02 crc kubenswrapper[4678]: healthz check failed Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.260870 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.295773 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.296766 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.306781 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.434842 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.435701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.435769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jp8j\" (UniqueName: \"kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.512019 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.517938 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd2343db-8449-4042-bf5e-b1441b226020" containerID="a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c" exitCode=0 Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.518002 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerDied","Data":"a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.521220 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3f0b571f87471a86c4871d4dd19b3d6f8452692357938cf53f1d3936b17e2ada"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.529374 4678 generic.go:334] "Generic (PLEG): container finished" podID="3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" containerID="14b85184f2bdeb4ab78a1e557551ba79024f2ecfd49b7bfedc0b91cf79029e95" exitCode=0 Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.529523 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7","Type":"ContainerDied","Data":"14b85184f2bdeb4ab78a1e557551ba79024f2ecfd49b7bfedc0b91cf79029e95"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.537482 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jp8j\" (UniqueName: \"kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.537528 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.537567 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.537928 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.538282 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.542027 4678 generic.go:334] "Generic (PLEG): container finished" podID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerID="989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04" exitCode=0 Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.542168 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerDied","Data":"989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.544926 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.548910 4678 generic.go:334] "Generic (PLEG): container finished" podID="e30f6aef-1479-4509-9515-f65af72e1c33" containerID="69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70" exitCode=0 Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.548987 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerDied","Data":"69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.566689 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jp8j\" (UniqueName: \"kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j\") pod \"redhat-marketplace-zhc66\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.568322 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" event={"ID":"56d1be59-35a1-44ac-873a-d5bc3df2f9ab","Type":"ContainerStarted","Data":"4aa6d6905e0ecf27a24a32fad81ed0d86c987b5721214e50ec0088149d88034a"} Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.617685 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.618025 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.632225 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vjqg7" podStartSLOduration=11.632208892 podStartE2EDuration="11.632208892s" podCreationTimestamp="2025-10-13 12:45:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:46:02.630727283 +0000 UTC m=+150.715265167" watchObservedRunningTime="2025-10-13 12:46:02.632208892 +0000 UTC m=+150.716746776" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.803574 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.818353 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-shbr8" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.903763 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.906543 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:46:02 crc kubenswrapper[4678]: E1013 12:46:02.906755 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="839c0ab6-3764-4934-9d39-a09227958854" containerName="collect-profiles" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.906772 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="839c0ab6-3764-4934-9d39-a09227958854" containerName="collect-profiles" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.906885 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="839c0ab6-3764-4934-9d39-a09227958854" containerName="collect-profiles" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.907599 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:02 crc kubenswrapper[4678]: W1013 12:46:02.914440 4678 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 13 12:46:02 crc kubenswrapper[4678]: E1013 12:46:02.914483 4678 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 13 12:46:02 crc kubenswrapper[4678]: I1013 12:46:02.953580 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.005172 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.070625 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume\") pod \"839c0ab6-3764-4934-9d39-a09227958854\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.070701 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sd69\" (UniqueName: \"kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69\") pod \"839c0ab6-3764-4934-9d39-a09227958854\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.070771 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume\") pod \"839c0ab6-3764-4934-9d39-a09227958854\" (UID: \"839c0ab6-3764-4934-9d39-a09227958854\") " Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.070968 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.071013 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.071041 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnjjj\" (UniqueName: \"kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.072111 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume" (OuterVolumeSpecName: "config-volume") pod "839c0ab6-3764-4934-9d39-a09227958854" (UID: "839c0ab6-3764-4934-9d39-a09227958854"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.084337 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "839c0ab6-3764-4934-9d39-a09227958854" (UID: "839c0ab6-3764-4934-9d39-a09227958854"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.085316 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69" (OuterVolumeSpecName: "kube-api-access-5sd69") pod "839c0ab6-3764-4934-9d39-a09227958854" (UID: "839c0ab6-3764-4934-9d39-a09227958854"). InnerVolumeSpecName "kube-api-access-5sd69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172182 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172212 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnjjj\" (UniqueName: \"kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172281 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/839c0ab6-3764-4934-9d39-a09227958854-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172295 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/839c0ab6-3764-4934-9d39-a09227958854-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172321 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sd69\" (UniqueName: \"kubernetes.io/projected/839c0ab6-3764-4934-9d39-a09227958854-kube-api-access-5sd69\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172614 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.172756 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.193407 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnjjj\" (UniqueName: \"kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj\") pod \"redhat-operators-qcptp\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.253687 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:46:03 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:46:03 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:46:03 crc kubenswrapper[4678]: healthz check failed Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.253777 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.295962 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.297018 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.315559 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.377986 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnvzs\" (UniqueName: \"kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.378104 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.378144 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.479090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnvzs\" (UniqueName: \"kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.479178 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.479209 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.479864 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.479947 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.497934 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnvzs\" (UniqueName: \"kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs\") pod \"redhat-operators-pfqsm\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.577902 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" event={"ID":"839c0ab6-3764-4934-9d39-a09227958854","Type":"ContainerDied","Data":"fcac5aac0b5d0af0677d288334e27317db41e45b4d6c2620062e9ba4c7160356"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.577944 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcac5aac0b5d0af0677d288334e27317db41e45b4d6c2620062e9ba4c7160356" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.578020 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.587822 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b962666-d096-465e-9dbb-6817a8b739bd" containerID="f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd" exitCode=0 Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.587892 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerDied","Data":"f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.587918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerStarted","Data":"123f4950053459da0f35df62b024b33020c40573e9cf88c42a3abd08b012adfc"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.607326 4678 generic.go:334] "Generic (PLEG): container finished" podID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerID="55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6" exitCode=0 Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.607840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerDied","Data":"55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.607880 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerStarted","Data":"8e1c1506441ecd042537f22cd48edba18eb8bbc20c359709c20d6c72d7c7cea6"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.618558 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" event={"ID":"afd9036b-48b5-478a-88e9-d5904ab84534","Type":"ContainerStarted","Data":"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.618612 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.618624 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" event={"ID":"afd9036b-48b5-478a-88e9-d5904ab84534","Type":"ContainerStarted","Data":"cb30b78a8737dfb89eb5b28d0bf0ca4fe9d05c4e36e04a0666dcb7b13e3b9f1d"} Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.680781 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" podStartSLOduration=130.680757485 podStartE2EDuration="2m10.680757485s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:46:03.661226042 +0000 UTC m=+151.745763926" watchObservedRunningTime="2025-10-13 12:46:03.680757485 +0000 UTC m=+151.765295369" Oct 13 12:46:03 crc kubenswrapper[4678]: E1013 12:46:03.698786 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod839c0ab6_3764_4934_9d39_a09227958854.slice/crio-fcac5aac0b5d0af0677d288334e27317db41e45b4d6c2620062e9ba4c7160356\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod839c0ab6_3764_4934_9d39_a09227958854.slice\": RecentStats: unable to find data in memory cache]" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.865186 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.989877 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir\") pod \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.989975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access\") pod \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\" (UID: \"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7\") " Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.990028 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" (UID: "3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.990437 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:03 crc kubenswrapper[4678]: I1013 12:46:03.996031 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" (UID: "3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.053751 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 12:46:04 crc kubenswrapper[4678]: E1013 12:46:04.053971 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" containerName="pruner" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.053983 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" containerName="pruner" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.054160 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7" containerName="pruner" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.055389 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.058544 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.058837 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.063222 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.091618 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.192644 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.193168 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.217906 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.218926 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.225614 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.259503 4678 patch_prober.go:28] interesting pod/router-default-5444994796-tsjvf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 13 12:46:04 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Oct 13 12:46:04 crc kubenswrapper[4678]: [+]process-running ok Oct 13 12:46:04 crc kubenswrapper[4678]: healthz check failed Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.259563 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tsjvf" podUID="6e30bbf0-723d-4233-b05f-a679c0e32ad5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.296010 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.296082 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.296190 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.327116 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.384686 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.577225 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-c9rl2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.577612 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c9rl2" podUID="f35010e9-46ee-4ba5-8e85-b7568bf88650" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.577380 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-c9rl2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.578142 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-c9rl2" podUID="f35010e9-46ee-4ba5-8e85-b7568bf88650" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.629762 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.630430 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3fe72e98-2fda-4aed-ba79-b4da1ea5bfd7","Type":"ContainerDied","Data":"94800d6b24d17822281ade95dfe6152e26d90f988e37759a6214454b9d0e1dc8"} Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.630469 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94800d6b24d17822281ade95dfe6152e26d90f988e37759a6214454b9d0e1dc8" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.644704 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.645956 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.648267 4678 patch_prober.go:28] interesting pod/console-f9d7485db-qq6wd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.648336 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qq6wd" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.765801 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:04 crc kubenswrapper[4678]: W1013 12:46:04.783828 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9fed871_b48d_4a79_b23b_77e758132e7e.slice/crio-9942cdf754192f927f61fd6f2dffcc6d6a358aae678869975a1b72c87e11a7d6 WatchSource:0}: Error finding container 9942cdf754192f927f61fd6f2dffcc6d6a358aae678869975a1b72c87e11a7d6: Status 404 returned error can't find the container with id 9942cdf754192f927f61fd6f2dffcc6d6a358aae678869975a1b72c87e11a7d6 Oct 13 12:46:04 crc kubenswrapper[4678]: I1013 12:46:04.873654 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.083460 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 13 12:46:05 crc kubenswrapper[4678]: W1013 12:46:05.228435 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod95719d97_b10d_4d43_a92d_d969ad6105eb.slice/crio-d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50 WatchSource:0}: Error finding container d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50: Status 404 returned error can't find the container with id d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50 Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.248998 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.261557 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.652505 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"95719d97-b10d-4d43-a92d-d969ad6105eb","Type":"ContainerStarted","Data":"d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50"} Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.700147 4678 generic.go:334] "Generic (PLEG): container finished" podID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerID="d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a" exitCode=0 Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.700444 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerDied","Data":"d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a"} Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.700794 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerStarted","Data":"ca327b173d81df9904a41a591ec46d8fb34f47350845e8ff2b0c8d822e25789e"} Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.711493 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerID="9e3f3cd87bbb6bd83631f39bec17058e09e7f10120e10b1d9946769a52237028" exitCode=0 Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.713130 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerDied","Data":"9e3f3cd87bbb6bd83631f39bec17058e09e7f10120e10b1d9946769a52237028"} Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.713157 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerStarted","Data":"9942cdf754192f927f61fd6f2dffcc6d6a358aae678869975a1b72c87e11a7d6"} Oct 13 12:46:05 crc kubenswrapper[4678]: I1013 12:46:05.721120 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-tsjvf" Oct 13 12:46:06 crc kubenswrapper[4678]: I1013 12:46:06.730655 4678 generic.go:334] "Generic (PLEG): container finished" podID="95719d97-b10d-4d43-a92d-d969ad6105eb" containerID="eb74ab58bcbd121390d1178e347ade6a991cef6bf6497c166c1e0faedb6091c6" exitCode=0 Oct 13 12:46:06 crc kubenswrapper[4678]: I1013 12:46:06.731398 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"95719d97-b10d-4d43-a92d-d969ad6105eb","Type":"ContainerDied","Data":"eb74ab58bcbd121390d1178e347ade6a991cef6bf6497c166c1e0faedb6091c6"} Oct 13 12:46:07 crc kubenswrapper[4678]: I1013 12:46:07.070677 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hkh7k" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.072204 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.167389 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir\") pod \"95719d97-b10d-4d43-a92d-d969ad6105eb\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.167455 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access\") pod \"95719d97-b10d-4d43-a92d-d969ad6105eb\" (UID: \"95719d97-b10d-4d43-a92d-d969ad6105eb\") " Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.168779 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "95719d97-b10d-4d43-a92d-d969ad6105eb" (UID: "95719d97-b10d-4d43-a92d-d969ad6105eb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.174295 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "95719d97-b10d-4d43-a92d-d969ad6105eb" (UID: "95719d97-b10d-4d43-a92d-d969ad6105eb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.269231 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95719d97-b10d-4d43-a92d-d969ad6105eb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.269264 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95719d97-b10d-4d43-a92d-d969ad6105eb-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.761808 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"95719d97-b10d-4d43-a92d-d969ad6105eb","Type":"ContainerDied","Data":"d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50"} Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.761849 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2649deca9938eabf36422e2a248a8dedcc0f2cae9a9973c68f5dda64962da50" Oct 13 12:46:08 crc kubenswrapper[4678]: I1013 12:46:08.761936 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 13 12:46:10 crc kubenswrapper[4678]: I1013 12:46:10.116798 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:46:14 crc kubenswrapper[4678]: I1013 12:46:14.587350 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-c9rl2" Oct 13 12:46:14 crc kubenswrapper[4678]: I1013 12:46:14.647111 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:46:14 crc kubenswrapper[4678]: I1013 12:46:14.650070 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:46:15 crc kubenswrapper[4678]: I1013 12:46:15.986491 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:46:15 crc kubenswrapper[4678]: I1013 12:46:15.995449 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ccdd1a7-2f11-43fd-9369-5ca8380ef70a-metrics-certs\") pod \"network-metrics-daemon-hbs76\" (UID: \"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a\") " pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:46:16 crc kubenswrapper[4678]: I1013 12:46:16.214493 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbs76" Oct 13 12:46:22 crc kubenswrapper[4678]: I1013 12:46:22.087685 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:46:25 crc kubenswrapper[4678]: I1013 12:46:25.507077 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:46:25 crc kubenswrapper[4678]: I1013 12:46:25.508195 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.141687 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hbs76"] Oct 13 12:46:28 crc kubenswrapper[4678]: W1013 12:46:28.198028 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ccdd1a7_2f11_43fd_9369_5ca8380ef70a.slice/crio-7e9fc95a17af22ae205f819dad8c3a3c3fc88c398562fcc0ee7b48a323bcebc7 WatchSource:0}: Error finding container 7e9fc95a17af22ae205f819dad8c3a3c3fc88c398562fcc0ee7b48a323bcebc7: Status 404 returned error can't find the container with id 7e9fc95a17af22ae205f819dad8c3a3c3fc88c398562fcc0ee7b48a323bcebc7 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.890641 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerID="17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.890813 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerDied","Data":"17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.893985 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd2343db-8449-4042-bf5e-b1441b226020" containerID="ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.894113 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerDied","Data":"ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.897705 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbs76" event={"ID":"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a","Type":"ContainerStarted","Data":"20cdb1ef62296ce3d0f9b2f480e6e0ec62c933a393768866de3001ade60ca4ca"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.897727 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbs76" event={"ID":"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a","Type":"ContainerStarted","Data":"7e9fc95a17af22ae205f819dad8c3a3c3fc88c398562fcc0ee7b48a323bcebc7"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.900509 4678 generic.go:334] "Generic (PLEG): container finished" podID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerID="ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.900576 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerDied","Data":"ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.904705 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerID="c62d987449924490a75a26e7ff72cd83465b62c0069063b3dbb06b1963205846" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.904749 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerDied","Data":"c62d987449924490a75a26e7ff72cd83465b62c0069063b3dbb06b1963205846"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.909252 4678 generic.go:334] "Generic (PLEG): container finished" podID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerID="089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.909439 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerDied","Data":"089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.915914 4678 generic.go:334] "Generic (PLEG): container finished" podID="e30f6aef-1479-4509-9515-f65af72e1c33" containerID="64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.918277 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerDied","Data":"64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.935147 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b962666-d096-465e-9dbb-6817a8b739bd" containerID="96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.935236 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerDied","Data":"96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b"} Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.938661 4678 generic.go:334] "Generic (PLEG): container finished" podID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerID="beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d" exitCode=0 Oct 13 12:46:28 crc kubenswrapper[4678]: I1013 12:46:28.938700 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerDied","Data":"beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d"} Oct 13 12:46:29 crc kubenswrapper[4678]: I1013 12:46:29.946677 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbs76" event={"ID":"4ccdd1a7-2f11-43fd-9369-5ca8380ef70a","Type":"ContainerStarted","Data":"8f38b12006cb37f7b9845b0a641fe73c775c312f3d4721a76ebe1cecbffd9832"} Oct 13 12:46:29 crc kubenswrapper[4678]: I1013 12:46:29.966539 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hbs76" podStartSLOduration=156.966523441 podStartE2EDuration="2m36.966523441s" podCreationTimestamp="2025-10-13 12:43:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:46:29.965105904 +0000 UTC m=+178.049643798" watchObservedRunningTime="2025-10-13 12:46:29.966523441 +0000 UTC m=+178.051061325" Oct 13 12:46:31 crc kubenswrapper[4678]: I1013 12:46:31.977349 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerStarted","Data":"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6"} Oct 13 12:46:32 crc kubenswrapper[4678]: I1013 12:46:32.009718 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p5f6j" podStartSLOduration=3.060264292 podStartE2EDuration="32.00968541s" podCreationTimestamp="2025-10-13 12:46:00 +0000 UTC" firstStartedPulling="2025-10-13 12:46:02.5621929 +0000 UTC m=+150.646730784" lastFinishedPulling="2025-10-13 12:46:31.511613988 +0000 UTC m=+179.596151902" observedRunningTime="2025-10-13 12:46:32.00511346 +0000 UTC m=+180.089651434" watchObservedRunningTime="2025-10-13 12:46:32.00968541 +0000 UTC m=+180.094223324" Oct 13 12:46:32 crc kubenswrapper[4678]: I1013 12:46:32.988666 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerStarted","Data":"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b"} Oct 13 12:46:32 crc kubenswrapper[4678]: I1013 12:46:32.994593 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerStarted","Data":"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679"} Oct 13 12:46:33 crc kubenswrapper[4678]: I1013 12:46:33.015707 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerStarted","Data":"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70"} Oct 13 12:46:33 crc kubenswrapper[4678]: I1013 12:46:33.022388 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qcptp" podStartSLOduration=4.148904073 podStartE2EDuration="31.02236568s" podCreationTimestamp="2025-10-13 12:46:02 +0000 UTC" firstStartedPulling="2025-10-13 12:46:05.704912014 +0000 UTC m=+153.789449898" lastFinishedPulling="2025-10-13 12:46:32.578373581 +0000 UTC m=+180.662911505" observedRunningTime="2025-10-13 12:46:33.007745945 +0000 UTC m=+181.092283829" watchObservedRunningTime="2025-10-13 12:46:33.02236568 +0000 UTC m=+181.106903574" Oct 13 12:46:33 crc kubenswrapper[4678]: I1013 12:46:33.054469 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2rk4n" podStartSLOduration=2.929997515 podStartE2EDuration="33.054448454s" podCreationTimestamp="2025-10-13 12:46:00 +0000 UTC" firstStartedPulling="2025-10-13 12:46:02.520551844 +0000 UTC m=+150.605089728" lastFinishedPulling="2025-10-13 12:46:32.645002773 +0000 UTC m=+180.729540667" observedRunningTime="2025-10-13 12:46:33.050064318 +0000 UTC m=+181.134602202" watchObservedRunningTime="2025-10-13 12:46:33.054448454 +0000 UTC m=+181.138986348" Oct 13 12:46:33 crc kubenswrapper[4678]: I1013 12:46:33.069257 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zhc66" podStartSLOduration=1.9950668139999999 podStartE2EDuration="31.069238843s" podCreationTimestamp="2025-10-13 12:46:02 +0000 UTC" firstStartedPulling="2025-10-13 12:46:03.592611837 +0000 UTC m=+151.677149721" lastFinishedPulling="2025-10-13 12:46:32.666783866 +0000 UTC m=+180.751321750" observedRunningTime="2025-10-13 12:46:33.065996317 +0000 UTC m=+181.150534201" watchObservedRunningTime="2025-10-13 12:46:33.069238843 +0000 UTC m=+181.153776727" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.023098 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerStarted","Data":"ba870dd3b9ff58d7512ce9e23bb4306cf2e05af619a1a791a415dad74faf6fbd"} Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.026419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerStarted","Data":"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673"} Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.029097 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerStarted","Data":"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f"} Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.031770 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerStarted","Data":"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645"} Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.067001 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9vhbd" podStartSLOduration=3.793872947 podStartE2EDuration="33.06698626s" podCreationTimestamp="2025-10-13 12:46:01 +0000 UTC" firstStartedPulling="2025-10-13 12:46:03.609253134 +0000 UTC m=+151.693791018" lastFinishedPulling="2025-10-13 12:46:32.882366437 +0000 UTC m=+180.966904331" observedRunningTime="2025-10-13 12:46:34.065451359 +0000 UTC m=+182.149989243" watchObservedRunningTime="2025-10-13 12:46:34.06698626 +0000 UTC m=+182.151524144" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.069176 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pfqsm" podStartSLOduration=4.032257811 podStartE2EDuration="31.069169027s" podCreationTimestamp="2025-10-13 12:46:03 +0000 UTC" firstStartedPulling="2025-10-13 12:46:05.716354385 +0000 UTC m=+153.800892269" lastFinishedPulling="2025-10-13 12:46:32.753265591 +0000 UTC m=+180.837803485" observedRunningTime="2025-10-13 12:46:34.047795805 +0000 UTC m=+182.132333699" watchObservedRunningTime="2025-10-13 12:46:34.069169027 +0000 UTC m=+182.153706901" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.085272 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pcx8f" podStartSLOduration=4.765226747 podStartE2EDuration="35.08525252s" podCreationTimestamp="2025-10-13 12:45:59 +0000 UTC" firstStartedPulling="2025-10-13 12:46:02.562241531 +0000 UTC m=+150.646779415" lastFinishedPulling="2025-10-13 12:46:32.882267294 +0000 UTC m=+180.966805188" observedRunningTime="2025-10-13 12:46:34.084184182 +0000 UTC m=+182.168722076" watchObservedRunningTime="2025-10-13 12:46:34.08525252 +0000 UTC m=+182.169790414" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.106206 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7jh7p" podStartSLOduration=3.737543052 podStartE2EDuration="35.106188431s" podCreationTimestamp="2025-10-13 12:45:59 +0000 UTC" firstStartedPulling="2025-10-13 12:46:01.465034977 +0000 UTC m=+149.549572861" lastFinishedPulling="2025-10-13 12:46:32.833680316 +0000 UTC m=+180.918218240" observedRunningTime="2025-10-13 12:46:34.103131381 +0000 UTC m=+182.187669265" watchObservedRunningTime="2025-10-13 12:46:34.106188431 +0000 UTC m=+182.190726315" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.219906 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.220002 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.226358 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:34 crc kubenswrapper[4678]: I1013 12:46:34.226409 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:35 crc kubenswrapper[4678]: I1013 12:46:35.338168 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2slxb" Oct 13 12:46:35 crc kubenswrapper[4678]: I1013 12:46:35.340063 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pfqsm" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="registry-server" probeResult="failure" output=< Oct 13 12:46:35 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Oct 13 12:46:35 crc kubenswrapper[4678]: > Oct 13 12:46:35 crc kubenswrapper[4678]: I1013 12:46:35.340862 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qcptp" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="registry-server" probeResult="failure" output=< Oct 13 12:46:35 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Oct 13 12:46:35 crc kubenswrapper[4678]: > Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.060202 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.061254 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.241803 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.241885 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.271522 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.282345 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.362236 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.549506 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.549565 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.606191 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.655473 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.655547 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:40 crc kubenswrapper[4678]: I1013 12:46:40.715719 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:41 crc kubenswrapper[4678]: I1013 12:46:41.115995 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:41 crc kubenswrapper[4678]: I1013 12:46:41.118721 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:46:41 crc kubenswrapper[4678]: I1013 12:46:41.134736 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:41 crc kubenswrapper[4678]: I1013 12:46:41.150557 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.232140 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.232541 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.273100 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.618542 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.619384 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.692301 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:42 crc kubenswrapper[4678]: I1013 12:46:42.936828 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.082936 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p5f6j" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="registry-server" containerID="cri-o://b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6" gracePeriod=2 Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.134392 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.143248 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.419250 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.531323 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.531622 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2rk4n" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="registry-server" containerID="cri-o://62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679" gracePeriod=2 Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.603979 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkxlq\" (UniqueName: \"kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq\") pod \"6eb35e8c-500e-43af-84f3-cf8ad776c127\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.604866 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content\") pod \"6eb35e8c-500e-43af-84f3-cf8ad776c127\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.604966 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities\") pod \"6eb35e8c-500e-43af-84f3-cf8ad776c127\" (UID: \"6eb35e8c-500e-43af-84f3-cf8ad776c127\") " Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.606150 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities" (OuterVolumeSpecName: "utilities") pod "6eb35e8c-500e-43af-84f3-cf8ad776c127" (UID: "6eb35e8c-500e-43af-84f3-cf8ad776c127"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.611185 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq" (OuterVolumeSpecName: "kube-api-access-xkxlq") pod "6eb35e8c-500e-43af-84f3-cf8ad776c127" (UID: "6eb35e8c-500e-43af-84f3-cf8ad776c127"). InnerVolumeSpecName "kube-api-access-xkxlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.661471 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6eb35e8c-500e-43af-84f3-cf8ad776c127" (UID: "6eb35e8c-500e-43af-84f3-cf8ad776c127"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.707020 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.707127 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eb35e8c-500e-43af-84f3-cf8ad776c127-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.707155 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkxlq\" (UniqueName: \"kubernetes.io/projected/6eb35e8c-500e-43af-84f3-cf8ad776c127-kube-api-access-xkxlq\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:43 crc kubenswrapper[4678]: I1013 12:46:43.861674 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.011007 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content\") pod \"fd2343db-8449-4042-bf5e-b1441b226020\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.011181 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtl79\" (UniqueName: \"kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79\") pod \"fd2343db-8449-4042-bf5e-b1441b226020\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.011250 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities\") pod \"fd2343db-8449-4042-bf5e-b1441b226020\" (UID: \"fd2343db-8449-4042-bf5e-b1441b226020\") " Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.012194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities" (OuterVolumeSpecName: "utilities") pod "fd2343db-8449-4042-bf5e-b1441b226020" (UID: "fd2343db-8449-4042-bf5e-b1441b226020"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.015541 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79" (OuterVolumeSpecName: "kube-api-access-qtl79") pod "fd2343db-8449-4042-bf5e-b1441b226020" (UID: "fd2343db-8449-4042-bf5e-b1441b226020"). InnerVolumeSpecName "kube-api-access-qtl79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.054262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd2343db-8449-4042-bf5e-b1441b226020" (UID: "fd2343db-8449-4042-bf5e-b1441b226020"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.090802 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd2343db-8449-4042-bf5e-b1441b226020" containerID="62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679" exitCode=0 Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.090858 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerDied","Data":"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679"} Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.090884 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2rk4n" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.090915 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2rk4n" event={"ID":"fd2343db-8449-4042-bf5e-b1441b226020","Type":"ContainerDied","Data":"6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390"} Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.090944 4678 scope.go:117] "RemoveContainer" containerID="62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.093394 4678 generic.go:334] "Generic (PLEG): container finished" podID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerID="b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6" exitCode=0 Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.093522 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerDied","Data":"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6"} Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.093647 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5f6j" event={"ID":"6eb35e8c-500e-43af-84f3-cf8ad776c127","Type":"ContainerDied","Data":"8159176dfeaf88a7365fd734a658927d211348e9a245f188786472bc3aad9ae2"} Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.093867 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5f6j" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.112486 4678 scope.go:117] "RemoveContainer" containerID="ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.112778 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtl79\" (UniqueName: \"kubernetes.io/projected/fd2343db-8449-4042-bf5e-b1441b226020-kube-api-access-qtl79\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.112805 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.112880 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2343db-8449-4042-bf5e-b1441b226020-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.133178 4678 scope.go:117] "RemoveContainer" containerID="a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.145564 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.153154 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2rk4n"] Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.159535 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.164700 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p5f6j"] Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.170805 4678 scope.go:117] "RemoveContainer" containerID="62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.171517 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679\": container with ID starting with 62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679 not found: ID does not exist" containerID="62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.171560 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679"} err="failed to get container status \"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679\": rpc error: code = NotFound desc = could not find container \"62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679\": container with ID starting with 62854acb7a78f931c5e10c487cbb328309c0f2c3b1efd46ef668cf3cd35e6679 not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.171611 4678 scope.go:117] "RemoveContainer" containerID="ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.172726 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc\": container with ID starting with ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc not found: ID does not exist" containerID="ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.172786 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc"} err="failed to get container status \"ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc\": rpc error: code = NotFound desc = could not find container \"ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc\": container with ID starting with ce2195ca0baa2c3c6af94b0a65a42d78994e7004debdb08314611aac0c5aa1cc not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.172807 4678 scope.go:117] "RemoveContainer" containerID="a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.173299 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c\": container with ID starting with a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c not found: ID does not exist" containerID="a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.173336 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c"} err="failed to get container status \"a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c\": rpc error: code = NotFound desc = could not find container \"a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c\": container with ID starting with a7c46b124223e071862413ad9ff3b93e953a34cb04888030291c5f3dd8a3d89c not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.173355 4678 scope.go:117] "RemoveContainer" containerID="b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.189286 4678 scope.go:117] "RemoveContainer" containerID="089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.196567 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd2343db_8449_4042_bf5e_b1441b226020.slice/crio-6b6c9c22ec171f6ae6377f337d17f6845c612f97f32b19336ad31b88a7f0c390\": RecentStats: unable to find data in memory cache]" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.207222 4678 scope.go:117] "RemoveContainer" containerID="989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.248319 4678 scope.go:117] "RemoveContainer" containerID="b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.249685 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6\": container with ID starting with b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6 not found: ID does not exist" containerID="b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.249745 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6"} err="failed to get container status \"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6\": rpc error: code = NotFound desc = could not find container \"b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6\": container with ID starting with b46ca489b8e9297519f11986305e636bbd84da71053e328be37ecb4bdaf695b6 not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.249787 4678 scope.go:117] "RemoveContainer" containerID="089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.250281 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492\": container with ID starting with 089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492 not found: ID does not exist" containerID="089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.250363 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492"} err="failed to get container status \"089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492\": rpc error: code = NotFound desc = could not find container \"089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492\": container with ID starting with 089d5b0f1c14fe2bbae4b50bcaa6e665cef66febe995e7af4e8f6f4bafd26492 not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.250395 4678 scope.go:117] "RemoveContainer" containerID="989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04" Oct 13 12:46:44 crc kubenswrapper[4678]: E1013 12:46:44.250767 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04\": container with ID starting with 989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04 not found: ID does not exist" containerID="989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.250802 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04"} err="failed to get container status \"989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04\": rpc error: code = NotFound desc = could not find container \"989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04\": container with ID starting with 989c93645fae525b97719a1c146ca63b71792a6062cb6692ab5b147e50f48f04 not found: ID does not exist" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.284536 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.294312 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.333851 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.339861 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.607975 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" path="/var/lib/kubelet/pods/6eb35e8c-500e-43af-84f3-cf8ad776c127/volumes" Oct 13 12:46:44 crc kubenswrapper[4678]: I1013 12:46:44.609276 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2343db-8449-4042-bf5e-b1441b226020" path="/var/lib/kubelet/pods/fd2343db-8449-4042-bf5e-b1441b226020/volumes" Oct 13 12:46:45 crc kubenswrapper[4678]: I1013 12:46:45.329566 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.113943 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zhc66" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="registry-server" containerID="cri-o://32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70" gracePeriod=2 Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.549271 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.653723 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content\") pod \"3b962666-d096-465e-9dbb-6817a8b739bd\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.653866 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities\") pod \"3b962666-d096-465e-9dbb-6817a8b739bd\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.653957 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jp8j\" (UniqueName: \"kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j\") pod \"3b962666-d096-465e-9dbb-6817a8b739bd\" (UID: \"3b962666-d096-465e-9dbb-6817a8b739bd\") " Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.655572 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities" (OuterVolumeSpecName: "utilities") pod "3b962666-d096-465e-9dbb-6817a8b739bd" (UID: "3b962666-d096-465e-9dbb-6817a8b739bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.664340 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j" (OuterVolumeSpecName: "kube-api-access-8jp8j") pod "3b962666-d096-465e-9dbb-6817a8b739bd" (UID: "3b962666-d096-465e-9dbb-6817a8b739bd"). InnerVolumeSpecName "kube-api-access-8jp8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.672413 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b962666-d096-465e-9dbb-6817a8b739bd" (UID: "3b962666-d096-465e-9dbb-6817a8b739bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.757897 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jp8j\" (UniqueName: \"kubernetes.io/projected/3b962666-d096-465e-9dbb-6817a8b739bd-kube-api-access-8jp8j\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.757947 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:46 crc kubenswrapper[4678]: I1013 12:46:46.757965 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b962666-d096-465e-9dbb-6817a8b739bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.120652 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b962666-d096-465e-9dbb-6817a8b739bd" containerID="32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70" exitCode=0 Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.120695 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerDied","Data":"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70"} Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.120716 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhc66" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.120736 4678 scope.go:117] "RemoveContainer" containerID="32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.120724 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhc66" event={"ID":"3b962666-d096-465e-9dbb-6817a8b739bd","Type":"ContainerDied","Data":"123f4950053459da0f35df62b024b33020c40573e9cf88c42a3abd08b012adfc"} Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.162564 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.165693 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhc66"] Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.170640 4678 scope.go:117] "RemoveContainer" containerID="96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.188600 4678 scope.go:117] "RemoveContainer" containerID="f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.209748 4678 scope.go:117] "RemoveContainer" containerID="32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70" Oct 13 12:46:47 crc kubenswrapper[4678]: E1013 12:46:47.210282 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70\": container with ID starting with 32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70 not found: ID does not exist" containerID="32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.210328 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70"} err="failed to get container status \"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70\": rpc error: code = NotFound desc = could not find container \"32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70\": container with ID starting with 32a75628d3ca81a61fb949fb701cf8830cbbaf3215fe629163c8b7d8079a9c70 not found: ID does not exist" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.210357 4678 scope.go:117] "RemoveContainer" containerID="96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b" Oct 13 12:46:47 crc kubenswrapper[4678]: E1013 12:46:47.210933 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b\": container with ID starting with 96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b not found: ID does not exist" containerID="96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.210954 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b"} err="failed to get container status \"96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b\": rpc error: code = NotFound desc = could not find container \"96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b\": container with ID starting with 96f1547fa80ff9195851c172efc6ae9fe846ed62f1795fe85df280193eb1668b not found: ID does not exist" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.210971 4678 scope.go:117] "RemoveContainer" containerID="f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd" Oct 13 12:46:47 crc kubenswrapper[4678]: E1013 12:46:47.211445 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd\": container with ID starting with f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd not found: ID does not exist" containerID="f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.211485 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd"} err="failed to get container status \"f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd\": rpc error: code = NotFound desc = could not find container \"f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd\": container with ID starting with f2489f4f497b0d08c7bb3c8bbf1c881fc31a5e4c99d00778af5d9a5942555fcd not found: ID does not exist" Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.932488 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:47 crc kubenswrapper[4678]: I1013 12:46:47.933040 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pfqsm" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="registry-server" containerID="cri-o://ba870dd3b9ff58d7512ce9e23bb4306cf2e05af619a1a791a415dad74faf6fbd" gracePeriod=2 Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.130541 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerID="ba870dd3b9ff58d7512ce9e23bb4306cf2e05af619a1a791a415dad74faf6fbd" exitCode=0 Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.130577 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerDied","Data":"ba870dd3b9ff58d7512ce9e23bb4306cf2e05af619a1a791a415dad74faf6fbd"} Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.263134 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.375291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content\") pod \"d9fed871-b48d-4a79-b23b-77e758132e7e\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.375402 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnvzs\" (UniqueName: \"kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs\") pod \"d9fed871-b48d-4a79-b23b-77e758132e7e\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.376491 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities\") pod \"d9fed871-b48d-4a79-b23b-77e758132e7e\" (UID: \"d9fed871-b48d-4a79-b23b-77e758132e7e\") " Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.377894 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities" (OuterVolumeSpecName: "utilities") pod "d9fed871-b48d-4a79-b23b-77e758132e7e" (UID: "d9fed871-b48d-4a79-b23b-77e758132e7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.379672 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs" (OuterVolumeSpecName: "kube-api-access-wnvzs") pod "d9fed871-b48d-4a79-b23b-77e758132e7e" (UID: "d9fed871-b48d-4a79-b23b-77e758132e7e"). InnerVolumeSpecName "kube-api-access-wnvzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.450286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9fed871-b48d-4a79-b23b-77e758132e7e" (UID: "d9fed871-b48d-4a79-b23b-77e758132e7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.477551 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnvzs\" (UniqueName: \"kubernetes.io/projected/d9fed871-b48d-4a79-b23b-77e758132e7e-kube-api-access-wnvzs\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.477592 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.477608 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9fed871-b48d-4a79-b23b-77e758132e7e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:46:48 crc kubenswrapper[4678]: I1013 12:46:48.598230 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" path="/var/lib/kubelet/pods/3b962666-d096-465e-9dbb-6817a8b739bd/volumes" Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.140594 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pfqsm" event={"ID":"d9fed871-b48d-4a79-b23b-77e758132e7e","Type":"ContainerDied","Data":"9942cdf754192f927f61fd6f2dffcc6d6a358aae678869975a1b72c87e11a7d6"} Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.141037 4678 scope.go:117] "RemoveContainer" containerID="ba870dd3b9ff58d7512ce9e23bb4306cf2e05af619a1a791a415dad74faf6fbd" Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.140783 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pfqsm" Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.163934 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.169674 4678 scope.go:117] "RemoveContainer" containerID="c62d987449924490a75a26e7ff72cd83465b62c0069063b3dbb06b1963205846" Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.172982 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pfqsm"] Oct 13 12:46:49 crc kubenswrapper[4678]: I1013 12:46:49.190175 4678 scope.go:117] "RemoveContainer" containerID="9e3f3cd87bbb6bd83631f39bec17058e09e7f10120e10b1d9946769a52237028" Oct 13 12:46:50 crc kubenswrapper[4678]: I1013 12:46:50.598117 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" path="/var/lib/kubelet/pods/d9fed871-b48d-4a79-b23b-77e758132e7e/volumes" Oct 13 12:46:53 crc kubenswrapper[4678]: I1013 12:46:53.347706 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:46:55 crc kubenswrapper[4678]: I1013 12:46:55.505929 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:46:55 crc kubenswrapper[4678]: I1013 12:46:55.506397 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:46:55 crc kubenswrapper[4678]: I1013 12:46:55.506445 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:46:55 crc kubenswrapper[4678]: I1013 12:46:55.506922 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 12:46:55 crc kubenswrapper[4678]: I1013 12:46:55.506975 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c" gracePeriod=600 Oct 13 12:46:56 crc kubenswrapper[4678]: I1013 12:46:56.181404 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c"} Oct 13 12:46:56 crc kubenswrapper[4678]: I1013 12:46:56.181351 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c" exitCode=0 Oct 13 12:46:56 crc kubenswrapper[4678]: I1013 12:46:56.182368 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447"} Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.371157 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" podUID="297a1de0-f405-4338-aeab-1bf11a7afc8c" containerName="oauth-openshift" containerID="cri-o://89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7" gracePeriod=15 Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.846012 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.894931 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-9gg9b"] Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.895536 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.895686 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.895811 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.895931 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.896122 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.896887 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.897031 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.897207 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.897332 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.897452 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.897586 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.897696 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.897857 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.897984 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.898158 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.898285 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.898420 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.898538 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="extract-content" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.898666 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95719d97-b10d-4d43-a92d-d969ad6105eb" containerName="pruner" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.898826 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="95719d97-b10d-4d43-a92d-d969ad6105eb" containerName="pruner" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.898954 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.899275 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.899397 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.899505 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.899617 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.899742 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="extract-utilities" Oct 13 12:47:18 crc kubenswrapper[4678]: E1013 12:47:18.899860 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="297a1de0-f405-4338-aeab-1bf11a7afc8c" containerName="oauth-openshift" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.899982 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="297a1de0-f405-4338-aeab-1bf11a7afc8c" containerName="oauth-openshift" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.900328 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b962666-d096-465e-9dbb-6817a8b739bd" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.900488 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="95719d97-b10d-4d43-a92d-d969ad6105eb" containerName="pruner" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.900632 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="297a1de0-f405-4338-aeab-1bf11a7afc8c" containerName="oauth-openshift" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.901190 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2343db-8449-4042-bf5e-b1441b226020" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.901357 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb35e8c-500e-43af-84f3-cf8ad776c127" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.901485 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9fed871-b48d-4a79-b23b-77e758132e7e" containerName="registry-server" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.902194 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.912702 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-9gg9b"] Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914567 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914629 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914681 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914723 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh4z8\" (UniqueName: \"kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914823 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914890 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914943 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.914991 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915029 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915103 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915155 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915197 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915271 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.915307 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session\") pod \"297a1de0-f405-4338-aeab-1bf11a7afc8c\" (UID: \"297a1de0-f405-4338-aeab-1bf11a7afc8c\") " Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.917209 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.917452 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.918166 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.918527 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.919528 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.929882 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8" (OuterVolumeSpecName: "kube-api-access-vh4z8") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "kube-api-access-vh4z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.945274 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.945792 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.946231 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.946625 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.947215 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.947473 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.947657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:18 crc kubenswrapper[4678]: I1013 12:47:18.947911 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "297a1de0-f405-4338-aeab-1bf11a7afc8c" (UID: "297a1de0-f405-4338-aeab-1bf11a7afc8c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.016735 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.016802 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.016843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k97rb\" (UniqueName: \"kubernetes.io/projected/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-kube-api-access-k97rb\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.016910 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.016975 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017014 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017126 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017296 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017357 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-dir\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017394 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017430 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-policies\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017542 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017602 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017682 4678 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017707 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017729 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017749 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017770 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017792 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh4z8\" (UniqueName: \"kubernetes.io/projected/297a1de0-f405-4338-aeab-1bf11a7afc8c-kube-api-access-vh4z8\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017817 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017839 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017860 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017880 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017899 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017919 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017938 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/297a1de0-f405-4338-aeab-1bf11a7afc8c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.017958 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/297a1de0-f405-4338-aeab-1bf11a7afc8c-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119465 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119552 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119583 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119623 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119652 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-dir\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-policies\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119762 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119797 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119826 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119855 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k97rb\" (UniqueName: \"kubernetes.io/projected/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-kube-api-access-k97rb\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.119904 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.120771 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-dir\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.121091 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-service-ca\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.121505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-audit-policies\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.122354 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.123199 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.123769 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.125324 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.125471 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-session\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.125647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-router-certs\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.126320 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.126805 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.126954 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-error\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.127501 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-v4-0-config-user-template-login\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.146949 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k97rb\" (UniqueName: \"kubernetes.io/projected/5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2-kube-api-access-k97rb\") pod \"oauth-openshift-785f95f6b7-9gg9b\" (UID: \"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2\") " pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.275591 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.330155 4678 generic.go:334] "Generic (PLEG): container finished" podID="297a1de0-f405-4338-aeab-1bf11a7afc8c" containerID="89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7" exitCode=0 Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.330204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" event={"ID":"297a1de0-f405-4338-aeab-1bf11a7afc8c","Type":"ContainerDied","Data":"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7"} Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.330233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" event={"ID":"297a1de0-f405-4338-aeab-1bf11a7afc8c","Type":"ContainerDied","Data":"d099c7f7e1e662e9c28ab84ad15257d6d3893b32fad2e45350c68fb237cf79b4"} Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.330252 4678 scope.go:117] "RemoveContainer" containerID="89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.330353 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4ggx9" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.373083 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.377805 4678 scope.go:117] "RemoveContainer" containerID="89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7" Oct 13 12:47:19 crc kubenswrapper[4678]: E1013 12:47:19.379174 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7\": container with ID starting with 89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7 not found: ID does not exist" containerID="89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.379250 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7"} err="failed to get container status \"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7\": rpc error: code = NotFound desc = could not find container \"89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7\": container with ID starting with 89b83d6efbbb6f10b79b44152778b906c64220371879dbb4a99655fc1a49d0f7 not found: ID does not exist" Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.387351 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4ggx9"] Oct 13 12:47:19 crc kubenswrapper[4678]: I1013 12:47:19.521807 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-785f95f6b7-9gg9b"] Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.341442 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" event={"ID":"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2","Type":"ContainerStarted","Data":"3bbacf2712ef3313f396e1c971c9532cf4c79a6531dbb9eca3a9b116e0ded7f9"} Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.341520 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" event={"ID":"5ae47c46-1bb1-44e9-8cc8-d3af3f05b0d2","Type":"ContainerStarted","Data":"8bbeef22aa2deaa25c04ac1e7aa2ca58a5640ce8e2150e364cc96e2161a0e9a6"} Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.341894 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.373332 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" podStartSLOduration=27.37329935 podStartE2EDuration="27.37329935s" podCreationTimestamp="2025-10-13 12:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:47:20.372079378 +0000 UTC m=+228.456617302" watchObservedRunningTime="2025-10-13 12:47:20.37329935 +0000 UTC m=+228.457837274" Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.415261 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-785f95f6b7-9gg9b" Oct 13 12:47:20 crc kubenswrapper[4678]: I1013 12:47:20.599259 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="297a1de0-f405-4338-aeab-1bf11a7afc8c" path="/var/lib/kubelet/pods/297a1de0-f405-4338-aeab-1bf11a7afc8c/volumes" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.448550 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.451280 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7jh7p" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="registry-server" containerID="cri-o://39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f" gracePeriod=30 Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.459008 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.459289 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pcx8f" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="registry-server" containerID="cri-o://cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673" gracePeriod=30 Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.464912 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.465762 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" containerID="cri-o://b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14" gracePeriod=30 Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.476253 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.477094 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9vhbd" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="registry-server" containerID="cri-o://9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645" gracePeriod=30 Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.478851 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.479007 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qcptp" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="registry-server" containerID="cri-o://370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b" gracePeriod=30 Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.493519 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9j9mq"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.494256 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.504468 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9j9mq"] Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.592953 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2vtt\" (UniqueName: \"kubernetes.io/projected/2c14e376-1a85-43dc-a3d1-66e37a54fd22-kube-api-access-w2vtt\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.593002 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.593023 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.694298 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2vtt\" (UniqueName: \"kubernetes.io/projected/2c14e376-1a85-43dc-a3d1-66e37a54fd22-kube-api-access-w2vtt\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.694346 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.694366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.695973 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.702503 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2c14e376-1a85-43dc-a3d1-66e37a54fd22-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.710795 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2vtt\" (UniqueName: \"kubernetes.io/projected/2c14e376-1a85-43dc-a3d1-66e37a54fd22-kube-api-access-w2vtt\") pod \"marketplace-operator-79b997595-9j9mq\" (UID: \"2c14e376-1a85-43dc-a3d1-66e37a54fd22\") " pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.818117 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.929953 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.930452 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.932093 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.935432 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.942670 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997509 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities\") pod \"e30f6aef-1479-4509-9515-f65af72e1c33\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997562 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities\") pod \"7f371c07-f1e3-4088-b222-d29499bdcfdf\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997601 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content\") pod \"e30f6aef-1479-4509-9515-f65af72e1c33\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997626 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content\") pod \"7f371c07-f1e3-4088-b222-d29499bdcfdf\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997671 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkfng\" (UniqueName: \"kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng\") pod \"e30f6aef-1479-4509-9515-f65af72e1c33\" (UID: \"e30f6aef-1479-4509-9515-f65af72e1c33\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997689 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities\") pod \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnjjj\" (UniqueName: \"kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj\") pod \"7f371c07-f1e3-4088-b222-d29499bdcfdf\" (UID: \"7f371c07-f1e3-4088-b222-d29499bdcfdf\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997734 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nspzk\" (UniqueName: \"kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk\") pod \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.997766 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content\") pod \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\" (UID: \"7c6ecd9e-93fe-446e-9026-857a2fa59aa1\") " Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.998971 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities" (OuterVolumeSpecName: "utilities") pod "7f371c07-f1e3-4088-b222-d29499bdcfdf" (UID: "7f371c07-f1e3-4088-b222-d29499bdcfdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:36 crc kubenswrapper[4678]: I1013 12:47:36.998995 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities" (OuterVolumeSpecName: "utilities") pod "e30f6aef-1479-4509-9515-f65af72e1c33" (UID: "e30f6aef-1479-4509-9515-f65af72e1c33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.000112 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities" (OuterVolumeSpecName: "utilities") pod "7c6ecd9e-93fe-446e-9026-857a2fa59aa1" (UID: "7c6ecd9e-93fe-446e-9026-857a2fa59aa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.008089 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk" (OuterVolumeSpecName: "kube-api-access-nspzk") pod "7c6ecd9e-93fe-446e-9026-857a2fa59aa1" (UID: "7c6ecd9e-93fe-446e-9026-857a2fa59aa1"). InnerVolumeSpecName "kube-api-access-nspzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.017656 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj" (OuterVolumeSpecName: "kube-api-access-rnjjj") pod "7f371c07-f1e3-4088-b222-d29499bdcfdf" (UID: "7f371c07-f1e3-4088-b222-d29499bdcfdf"). InnerVolumeSpecName "kube-api-access-rnjjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.023359 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng" (OuterVolumeSpecName: "kube-api-access-bkfng") pod "e30f6aef-1479-4509-9515-f65af72e1c33" (UID: "e30f6aef-1479-4509-9515-f65af72e1c33"). InnerVolumeSpecName "kube-api-access-bkfng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.071138 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c6ecd9e-93fe-446e-9026-857a2fa59aa1" (UID: "7c6ecd9e-93fe-446e-9026-857a2fa59aa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.071992 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e30f6aef-1479-4509-9515-f65af72e1c33" (UID: "e30f6aef-1479-4509-9515-f65af72e1c33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.094300 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f371c07-f1e3-4088-b222-d29499bdcfdf" (UID: "7f371c07-f1e3-4088-b222-d29499bdcfdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098553 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca\") pod \"6b555cbc-dd4d-45bc-b311-ac341169716b\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098606 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content\") pod \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098645 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flgnf\" (UniqueName: \"kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf\") pod \"6b555cbc-dd4d-45bc-b311-ac341169716b\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098667 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xw9f\" (UniqueName: \"kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f\") pod \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098786 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics\") pod \"6b555cbc-dd4d-45bc-b311-ac341169716b\" (UID: \"6b555cbc-dd4d-45bc-b311-ac341169716b\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.098825 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities\") pod \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\" (UID: \"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f\") " Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099107 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkfng\" (UniqueName: \"kubernetes.io/projected/e30f6aef-1479-4509-9515-f65af72e1c33-kube-api-access-bkfng\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099130 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099141 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnjjj\" (UniqueName: \"kubernetes.io/projected/7f371c07-f1e3-4088-b222-d29499bdcfdf-kube-api-access-rnjjj\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099152 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nspzk\" (UniqueName: \"kubernetes.io/projected/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-kube-api-access-nspzk\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099162 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c6ecd9e-93fe-446e-9026-857a2fa59aa1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099171 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099179 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099188 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e30f6aef-1479-4509-9515-f65af72e1c33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099183 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6b555cbc-dd4d-45bc-b311-ac341169716b" (UID: "6b555cbc-dd4d-45bc-b311-ac341169716b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099198 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f371c07-f1e3-4088-b222-d29499bdcfdf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.099938 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities" (OuterVolumeSpecName: "utilities") pod "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" (UID: "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.102077 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf" (OuterVolumeSpecName: "kube-api-access-flgnf") pod "6b555cbc-dd4d-45bc-b311-ac341169716b" (UID: "6b555cbc-dd4d-45bc-b311-ac341169716b"). InnerVolumeSpecName "kube-api-access-flgnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.102303 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f" (OuterVolumeSpecName: "kube-api-access-2xw9f") pod "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" (UID: "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f"). InnerVolumeSpecName "kube-api-access-2xw9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.102531 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6b555cbc-dd4d-45bc-b311-ac341169716b" (UID: "6b555cbc-dd4d-45bc-b311-ac341169716b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.113787 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" (UID: "4cee0216-ff54-4cba-b85d-3ee7d5f5d22f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.199991 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.200024 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.200036 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b555cbc-dd4d-45bc-b311-ac341169716b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.200045 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.200071 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flgnf\" (UniqueName: \"kubernetes.io/projected/6b555cbc-dd4d-45bc-b311-ac341169716b-kube-api-access-flgnf\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.200080 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xw9f\" (UniqueName: \"kubernetes.io/projected/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f-kube-api-access-2xw9f\") on node \"crc\" DevicePath \"\"" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.249661 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9j9mq"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.454898 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" event={"ID":"2c14e376-1a85-43dc-a3d1-66e37a54fd22","Type":"ContainerStarted","Data":"f728bf790462d37776526b9b9c1cb439fcac5f996b84d7ef4001304b8c39feb8"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.456207 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" event={"ID":"2c14e376-1a85-43dc-a3d1-66e37a54fd22","Type":"ContainerStarted","Data":"6ede269d2a722f737ebab954aceec2492c95dc91205fefceca010df707bf4b43"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.456245 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.457869 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9j9mq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.457912 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" podUID="2c14e376-1a85-43dc-a3d1-66e37a54fd22" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.457971 4678 generic.go:334] "Generic (PLEG): container finished" podID="e30f6aef-1479-4509-9515-f65af72e1c33" containerID="cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673" exitCode=0 Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.458021 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerDied","Data":"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.458046 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcx8f" event={"ID":"e30f6aef-1479-4509-9515-f65af72e1c33","Type":"ContainerDied","Data":"c6c5409b0ac8be52cdc99978807ea5dcdfae9ee0cf833f8bf1c1dbafcbf2aead"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.458088 4678 scope.go:117] "RemoveContainer" containerID="cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.458152 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcx8f" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.460287 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerID="39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f" exitCode=0 Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.460331 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerDied","Data":"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.460348 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jh7p" event={"ID":"7c6ecd9e-93fe-446e-9026-857a2fa59aa1","Type":"ContainerDied","Data":"7edd5476c30dc3cc1212ae750d2b71acd6d8a42a795ec9d06d603cdb1e3aad7a"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.460357 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jh7p" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.465747 4678 generic.go:334] "Generic (PLEG): container finished" podID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerID="b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14" exitCode=0 Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.465801 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" event={"ID":"6b555cbc-dd4d-45bc-b311-ac341169716b","Type":"ContainerDied","Data":"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.465822 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" event={"ID":"6b555cbc-dd4d-45bc-b311-ac341169716b","Type":"ContainerDied","Data":"0a6756d33e03b850ad2a9b4ce6a15e1ee364122c9f71ca482027ff6428fc0c19"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.465890 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-pj7h9" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.468107 4678 generic.go:334] "Generic (PLEG): container finished" podID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerID="370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b" exitCode=0 Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.468161 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerDied","Data":"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.468182 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qcptp" event={"ID":"7f371c07-f1e3-4088-b222-d29499bdcfdf","Type":"ContainerDied","Data":"ca327b173d81df9904a41a591ec46d8fb34f47350845e8ff2b0c8d822e25789e"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.468244 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qcptp" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.470972 4678 generic.go:334] "Generic (PLEG): container finished" podID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerID="9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645" exitCode=0 Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.470996 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerDied","Data":"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.471012 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9vhbd" event={"ID":"4cee0216-ff54-4cba-b85d-3ee7d5f5d22f","Type":"ContainerDied","Data":"8e1c1506441ecd042537f22cd48edba18eb8bbc20c359709c20d6c72d7c7cea6"} Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.471065 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9vhbd" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.474414 4678 scope.go:117] "RemoveContainer" containerID="64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.482793 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" podStartSLOduration=1.482774693 podStartE2EDuration="1.482774693s" podCreationTimestamp="2025-10-13 12:47:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:47:37.476178539 +0000 UTC m=+245.560716423" watchObservedRunningTime="2025-10-13 12:47:37.482774693 +0000 UTC m=+245.567312587" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.502625 4678 scope.go:117] "RemoveContainer" containerID="69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.524608 4678 scope.go:117] "RemoveContainer" containerID="cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.525247 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.525675 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673\": container with ID starting with cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673 not found: ID does not exist" containerID="cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.525727 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673"} err="failed to get container status \"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673\": rpc error: code = NotFound desc = could not find container \"cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673\": container with ID starting with cbbc4380781831a5d39d91650adbff86d6c7741c881b59ba9e0e17516bbb9673 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.525762 4678 scope.go:117] "RemoveContainer" containerID="64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.526247 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a\": container with ID starting with 64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a not found: ID does not exist" containerID="64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.526279 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a"} err="failed to get container status \"64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a\": rpc error: code = NotFound desc = could not find container \"64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a\": container with ID starting with 64a275ba3d8163d18c168e6e9ab9bcad1d4f7da38f48f83acfb75eb5a6df323a not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.526300 4678 scope.go:117] "RemoveContainer" containerID="69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.526666 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70\": container with ID starting with 69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70 not found: ID does not exist" containerID="69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.526716 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70"} err="failed to get container status \"69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70\": rpc error: code = NotFound desc = could not find container \"69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70\": container with ID starting with 69313766693b200bdd3206156f87ab46b6625e15d9595b39403dc3d6689f1f70 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.526748 4678 scope.go:117] "RemoveContainer" containerID="39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.540301 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7jh7p"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.549352 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.549398 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pcx8f"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.552692 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.555044 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qcptp"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.561295 4678 scope.go:117] "RemoveContainer" containerID="17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.571848 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.576545 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9vhbd"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.580376 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.580524 4678 scope.go:117] "RemoveContainer" containerID="cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.583241 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-pj7h9"] Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.621414 4678 scope.go:117] "RemoveContainer" containerID="39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.621862 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f\": container with ID starting with 39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f not found: ID does not exist" containerID="39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.621890 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f"} err="failed to get container status \"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f\": rpc error: code = NotFound desc = could not find container \"39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f\": container with ID starting with 39812b56de1cf21b4b174d6940dde3a02e37a603f435963a44e1e31636dfb24f not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.621911 4678 scope.go:117] "RemoveContainer" containerID="17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.623026 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05\": container with ID starting with 17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05 not found: ID does not exist" containerID="17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.623078 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05"} err="failed to get container status \"17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05\": rpc error: code = NotFound desc = could not find container \"17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05\": container with ID starting with 17e9b59524d67877336cf1b4794548ceddcce46afcd740634a8f829383adbe05 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.623091 4678 scope.go:117] "RemoveContainer" containerID="cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.623739 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e\": container with ID starting with cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e not found: ID does not exist" containerID="cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.623761 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e"} err="failed to get container status \"cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e\": rpc error: code = NotFound desc = could not find container \"cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e\": container with ID starting with cc6d72e490fa703880febfbffcd7859e5c458be18a769be4a35116758e61114e not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.623775 4678 scope.go:117] "RemoveContainer" containerID="b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.639689 4678 scope.go:117] "RemoveContainer" containerID="b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.640085 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14\": container with ID starting with b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14 not found: ID does not exist" containerID="b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.640108 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14"} err="failed to get container status \"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14\": rpc error: code = NotFound desc = could not find container \"b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14\": container with ID starting with b1137bea8b9834e10b02fb1ee49f0dc11a96c65dd75e32be77630aa8c60fca14 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.640123 4678 scope.go:117] "RemoveContainer" containerID="370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.654574 4678 scope.go:117] "RemoveContainer" containerID="beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.671233 4678 scope.go:117] "RemoveContainer" containerID="d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.685253 4678 scope.go:117] "RemoveContainer" containerID="370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.685619 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b\": container with ID starting with 370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b not found: ID does not exist" containerID="370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.685647 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b"} err="failed to get container status \"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b\": rpc error: code = NotFound desc = could not find container \"370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b\": container with ID starting with 370e893484655d1ff3c6b57ccd405c6e95e58edfdcbd056c70e7fa846788514b not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.685673 4678 scope.go:117] "RemoveContainer" containerID="beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.685914 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d\": container with ID starting with beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d not found: ID does not exist" containerID="beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.685932 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d"} err="failed to get container status \"beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d\": rpc error: code = NotFound desc = could not find container \"beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d\": container with ID starting with beda7f750be68ded3de10c702181b8ea0c28d0a2f1796f4d7e01f5ff1b0de96d not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.685944 4678 scope.go:117] "RemoveContainer" containerID="d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.686142 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a\": container with ID starting with d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a not found: ID does not exist" containerID="d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.686156 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a"} err="failed to get container status \"d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a\": rpc error: code = NotFound desc = could not find container \"d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a\": container with ID starting with d29cd5a617a8febc1d34f68623823b3d36118295f4536a138f5e2d40eaf3b46a not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.686168 4678 scope.go:117] "RemoveContainer" containerID="9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.698909 4678 scope.go:117] "RemoveContainer" containerID="ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.715328 4678 scope.go:117] "RemoveContainer" containerID="55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.731277 4678 scope.go:117] "RemoveContainer" containerID="9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.731776 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645\": container with ID starting with 9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645 not found: ID does not exist" containerID="9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.731829 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645"} err="failed to get container status \"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645\": rpc error: code = NotFound desc = could not find container \"9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645\": container with ID starting with 9cbb7be580cc17fe48b9fca91b0d5f5705a17f410dab147ddb50c6020fa1a645 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.731875 4678 scope.go:117] "RemoveContainer" containerID="ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.732199 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26\": container with ID starting with ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26 not found: ID does not exist" containerID="ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.732229 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26"} err="failed to get container status \"ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26\": rpc error: code = NotFound desc = could not find container \"ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26\": container with ID starting with ec6fb360c5acf104d3cb342563e82e938be331be8acdd566057ae26162c67e26 not found: ID does not exist" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.732255 4678 scope.go:117] "RemoveContainer" containerID="55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6" Oct 13 12:47:37 crc kubenswrapper[4678]: E1013 12:47:37.732575 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6\": container with ID starting with 55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6 not found: ID does not exist" containerID="55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6" Oct 13 12:47:37 crc kubenswrapper[4678]: I1013 12:47:37.732615 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6"} err="failed to get container status \"55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6\": rpc error: code = NotFound desc = could not find container \"55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6\": container with ID starting with 55eb5a17546aadb46aa742b2042e40cd586aadc38849dfe50b88b43d111bdca6 not found: ID does not exist" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.484028 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9j9mq" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.599477 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" path="/var/lib/kubelet/pods/4cee0216-ff54-4cba-b85d-3ee7d5f5d22f/volumes" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.600380 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" path="/var/lib/kubelet/pods/6b555cbc-dd4d-45bc-b311-ac341169716b/volumes" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.600904 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" path="/var/lib/kubelet/pods/7c6ecd9e-93fe-446e-9026-857a2fa59aa1/volumes" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.601892 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" path="/var/lib/kubelet/pods/7f371c07-f1e3-4088-b222-d29499bdcfdf/volumes" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.602452 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" path="/var/lib/kubelet/pods/e30f6aef-1479-4509-9515-f65af72e1c33/volumes" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666276 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-brf7p"] Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666520 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666544 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666578 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666591 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666609 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666622 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666640 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666652 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666670 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666682 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666696 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666707 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666724 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666738 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666754 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666768 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="extract-utilities" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666788 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666800 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666820 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666832 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666843 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666855 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666872 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666884 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" Oct 13 12:47:38 crc kubenswrapper[4678]: E1013 12:47:38.666898 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.666911 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="extract-content" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.667147 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f371c07-f1e3-4088-b222-d29499bdcfdf" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.667168 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c6ecd9e-93fe-446e-9026-857a2fa59aa1" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.667187 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cee0216-ff54-4cba-b85d-3ee7d5f5d22f" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.667204 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e30f6aef-1479-4509-9515-f65af72e1c33" containerName="registry-server" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.667224 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b555cbc-dd4d-45bc-b311-ac341169716b" containerName="marketplace-operator" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.668299 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.671923 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.679761 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brf7p"] Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.819246 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v78lp\" (UniqueName: \"kubernetes.io/projected/7495eaa2-5ad8-4110-96a8-11d4eab8682f-kube-api-access-v78lp\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.819467 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-utilities\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.819517 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-catalog-content\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.869745 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-56qm8"] Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.870719 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.872646 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.911805 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-56qm8"] Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.921172 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-utilities\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.921266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-catalog-content\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.921453 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v78lp\" (UniqueName: \"kubernetes.io/projected/7495eaa2-5ad8-4110-96a8-11d4eab8682f-kube-api-access-v78lp\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.921876 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-utilities\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.921902 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7495eaa2-5ad8-4110-96a8-11d4eab8682f-catalog-content\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.945851 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v78lp\" (UniqueName: \"kubernetes.io/projected/7495eaa2-5ad8-4110-96a8-11d4eab8682f-kube-api-access-v78lp\") pod \"redhat-marketplace-brf7p\" (UID: \"7495eaa2-5ad8-4110-96a8-11d4eab8682f\") " pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:38 crc kubenswrapper[4678]: I1013 12:47:38.984537 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.024377 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-utilities\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.024418 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-catalog-content\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.024467 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkcvg\" (UniqueName: \"kubernetes.io/projected/2ff166cb-c0b3-4451-9666-9a27bf005741-kube-api-access-pkcvg\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.125649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkcvg\" (UniqueName: \"kubernetes.io/projected/2ff166cb-c0b3-4451-9666-9a27bf005741-kube-api-access-pkcvg\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.125735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-utilities\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.125779 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-catalog-content\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.126723 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-catalog-content\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.126767 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff166cb-c0b3-4451-9666-9a27bf005741-utilities\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.148825 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkcvg\" (UniqueName: \"kubernetes.io/projected/2ff166cb-c0b3-4451-9666-9a27bf005741-kube-api-access-pkcvg\") pod \"redhat-operators-56qm8\" (UID: \"2ff166cb-c0b3-4451-9666-9a27bf005741\") " pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.193905 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brf7p"] Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.222369 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.391591 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-56qm8"] Oct 13 12:47:39 crc kubenswrapper[4678]: W1013 12:47:39.462596 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ff166cb_c0b3_4451_9666_9a27bf005741.slice/crio-0c5378af2768a10a4528ed08c0a42c5a1ff0a288c3a0e47eab646a9379be70c0 WatchSource:0}: Error finding container 0c5378af2768a10a4528ed08c0a42c5a1ff0a288c3a0e47eab646a9379be70c0: Status 404 returned error can't find the container with id 0c5378af2768a10a4528ed08c0a42c5a1ff0a288c3a0e47eab646a9379be70c0 Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.487691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-56qm8" event={"ID":"2ff166cb-c0b3-4451-9666-9a27bf005741","Type":"ContainerStarted","Data":"0c5378af2768a10a4528ed08c0a42c5a1ff0a288c3a0e47eab646a9379be70c0"} Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.493539 4678 generic.go:334] "Generic (PLEG): container finished" podID="7495eaa2-5ad8-4110-96a8-11d4eab8682f" containerID="4f3fb2cedbc65f952a232a87a13ef59e8478ead55162f95a582997a80f37c238" exitCode=0 Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.494071 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brf7p" event={"ID":"7495eaa2-5ad8-4110-96a8-11d4eab8682f","Type":"ContainerDied","Data":"4f3fb2cedbc65f952a232a87a13ef59e8478ead55162f95a582997a80f37c238"} Oct 13 12:47:39 crc kubenswrapper[4678]: I1013 12:47:39.494155 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brf7p" event={"ID":"7495eaa2-5ad8-4110-96a8-11d4eab8682f","Type":"ContainerStarted","Data":"4c5f82e0a473f04ef41c111eefc9ca843cfaf476e8cec4ebf9d377a5fcca22de"} Oct 13 12:47:40 crc kubenswrapper[4678]: I1013 12:47:40.502340 4678 generic.go:334] "Generic (PLEG): container finished" podID="7495eaa2-5ad8-4110-96a8-11d4eab8682f" containerID="103b38796d7766e6bb30b7b80c10d717a4c000147991cde8d1bbfddeafa0c1ac" exitCode=0 Oct 13 12:47:40 crc kubenswrapper[4678]: I1013 12:47:40.502434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brf7p" event={"ID":"7495eaa2-5ad8-4110-96a8-11d4eab8682f","Type":"ContainerDied","Data":"103b38796d7766e6bb30b7b80c10d717a4c000147991cde8d1bbfddeafa0c1ac"} Oct 13 12:47:40 crc kubenswrapper[4678]: I1013 12:47:40.504763 4678 generic.go:334] "Generic (PLEG): container finished" podID="2ff166cb-c0b3-4451-9666-9a27bf005741" containerID="5b2d98ee98d34be0143b5abf4524a18d0b6930041aa93a6d5f05df0fbb52badc" exitCode=0 Oct 13 12:47:40 crc kubenswrapper[4678]: I1013 12:47:40.505148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-56qm8" event={"ID":"2ff166cb-c0b3-4451-9666-9a27bf005741","Type":"ContainerDied","Data":"5b2d98ee98d34be0143b5abf4524a18d0b6930041aa93a6d5f05df0fbb52badc"} Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.075296 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b2nmp"] Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.076607 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.080503 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.089168 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b2nmp"] Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.151249 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-utilities\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.151425 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-catalog-content\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.151485 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr7ls\" (UniqueName: \"kubernetes.io/projected/ab56d409-9f69-413f-b208-d236ff08ed98-kube-api-access-cr7ls\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.252208 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-utilities\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.252283 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-catalog-content\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.252308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr7ls\" (UniqueName: \"kubernetes.io/projected/ab56d409-9f69-413f-b208-d236ff08ed98-kube-api-access-cr7ls\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.252905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-utilities\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.253170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab56d409-9f69-413f-b208-d236ff08ed98-catalog-content\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.273117 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.274700 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.276846 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.287771 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.295192 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr7ls\" (UniqueName: \"kubernetes.io/projected/ab56d409-9f69-413f-b208-d236ff08ed98-kube-api-access-cr7ls\") pod \"certified-operators-b2nmp\" (UID: \"ab56d409-9f69-413f-b208-d236ff08ed98\") " pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.353629 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq6qc\" (UniqueName: \"kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.353713 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.353757 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.404115 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.454680 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq6qc\" (UniqueName: \"kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.455242 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.455342 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.456010 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.456607 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.471936 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq6qc\" (UniqueName: \"kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc\") pod \"community-operators-hkpp7\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.513581 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brf7p" event={"ID":"7495eaa2-5ad8-4110-96a8-11d4eab8682f","Type":"ContainerStarted","Data":"5a975dd146bb3ae07e33084adb2722fa81da51d553c0bcbfb183d136447be009"} Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.539710 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-brf7p" podStartSLOduration=2.119696233 podStartE2EDuration="3.539689817s" podCreationTimestamp="2025-10-13 12:47:38 +0000 UTC" firstStartedPulling="2025-10-13 12:47:39.495645558 +0000 UTC m=+247.580183442" lastFinishedPulling="2025-10-13 12:47:40.915639142 +0000 UTC m=+249.000177026" observedRunningTime="2025-10-13 12:47:41.538730831 +0000 UTC m=+249.623268715" watchObservedRunningTime="2025-10-13 12:47:41.539689817 +0000 UTC m=+249.624227701" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.593870 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.639086 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b2nmp"] Oct 13 12:47:41 crc kubenswrapper[4678]: W1013 12:47:41.679037 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab56d409_9f69_413f_b208_d236ff08ed98.slice/crio-08b64c4ae6a6845c572a1e2a562382f7d2df4f87b18e8390519229c8d798bf44 WatchSource:0}: Error finding container 08b64c4ae6a6845c572a1e2a562382f7d2df4f87b18e8390519229c8d798bf44: Status 404 returned error can't find the container with id 08b64c4ae6a6845c572a1e2a562382f7d2df4f87b18e8390519229c8d798bf44 Oct 13 12:47:41 crc kubenswrapper[4678]: I1013 12:47:41.822083 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 12:47:41 crc kubenswrapper[4678]: W1013 12:47:41.873726 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4be8bfa_2d9f_490b_b239_852793e85b6b.slice/crio-cb5975440ede515e9c46bce351a488d6ef761fbd56e24cc97cc754a392b283a4 WatchSource:0}: Error finding container cb5975440ede515e9c46bce351a488d6ef761fbd56e24cc97cc754a392b283a4: Status 404 returned error can't find the container with id cb5975440ede515e9c46bce351a488d6ef761fbd56e24cc97cc754a392b283a4 Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.529833 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerID="b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5" exitCode=0 Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.529917 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerDied","Data":"b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5"} Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.529944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerStarted","Data":"cb5975440ede515e9c46bce351a488d6ef761fbd56e24cc97cc754a392b283a4"} Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.532093 4678 generic.go:334] "Generic (PLEG): container finished" podID="ab56d409-9f69-413f-b208-d236ff08ed98" containerID="2e45445657c411df16dc6621a39362807a46f73cd23da5267c77428219fb5888" exitCode=0 Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.532135 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2nmp" event={"ID":"ab56d409-9f69-413f-b208-d236ff08ed98","Type":"ContainerDied","Data":"2e45445657c411df16dc6621a39362807a46f73cd23da5267c77428219fb5888"} Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.532149 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2nmp" event={"ID":"ab56d409-9f69-413f-b208-d236ff08ed98","Type":"ContainerStarted","Data":"08b64c4ae6a6845c572a1e2a562382f7d2df4f87b18e8390519229c8d798bf44"} Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.536625 4678 generic.go:334] "Generic (PLEG): container finished" podID="2ff166cb-c0b3-4451-9666-9a27bf005741" containerID="e4ce3378d2cc7c706326da3f5daa027db68d61e0448c70d74fa58eadea4ccd2a" exitCode=0 Oct 13 12:47:42 crc kubenswrapper[4678]: I1013 12:47:42.536815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-56qm8" event={"ID":"2ff166cb-c0b3-4451-9666-9a27bf005741","Type":"ContainerDied","Data":"e4ce3378d2cc7c706326da3f5daa027db68d61e0448c70d74fa58eadea4ccd2a"} Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.543173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-56qm8" event={"ID":"2ff166cb-c0b3-4451-9666-9a27bf005741","Type":"ContainerStarted","Data":"dc365153e0008d2718cd1b474227102174b8384b3771bb4728e98bbb89342fa0"} Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.546125 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerID="75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f" exitCode=0 Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.546203 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerDied","Data":"75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f"} Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.547874 4678 generic.go:334] "Generic (PLEG): container finished" podID="ab56d409-9f69-413f-b208-d236ff08ed98" containerID="ad0388470a22c044013bcfe34e39d031e4d197e94145c1101e1f8d0c6d26e7cb" exitCode=0 Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.547897 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2nmp" event={"ID":"ab56d409-9f69-413f-b208-d236ff08ed98","Type":"ContainerDied","Data":"ad0388470a22c044013bcfe34e39d031e4d197e94145c1101e1f8d0c6d26e7cb"} Oct 13 12:47:43 crc kubenswrapper[4678]: I1013 12:47:43.563813 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-56qm8" podStartSLOduration=3.028750905 podStartE2EDuration="5.563795319s" podCreationTimestamp="2025-10-13 12:47:38 +0000 UTC" firstStartedPulling="2025-10-13 12:47:40.506595305 +0000 UTC m=+248.591133199" lastFinishedPulling="2025-10-13 12:47:43.041639729 +0000 UTC m=+251.126177613" observedRunningTime="2025-10-13 12:47:43.560661126 +0000 UTC m=+251.645199010" watchObservedRunningTime="2025-10-13 12:47:43.563795319 +0000 UTC m=+251.648333203" Oct 13 12:47:45 crc kubenswrapper[4678]: I1013 12:47:45.561338 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b2nmp" event={"ID":"ab56d409-9f69-413f-b208-d236ff08ed98","Type":"ContainerStarted","Data":"6b83b10beb56848f48d56ac80f2cb0772f8bbb836a9930970e089a0060db4cea"} Oct 13 12:47:45 crc kubenswrapper[4678]: I1013 12:47:45.573834 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerStarted","Data":"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc"} Oct 13 12:47:45 crc kubenswrapper[4678]: I1013 12:47:45.589913 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b2nmp" podStartSLOduration=3.119977078 podStartE2EDuration="4.589894423s" podCreationTimestamp="2025-10-13 12:47:41 +0000 UTC" firstStartedPulling="2025-10-13 12:47:42.535672388 +0000 UTC m=+250.620210302" lastFinishedPulling="2025-10-13 12:47:44.005589773 +0000 UTC m=+252.090127647" observedRunningTime="2025-10-13 12:47:45.58867271 +0000 UTC m=+253.673210634" watchObservedRunningTime="2025-10-13 12:47:45.589894423 +0000 UTC m=+253.674432307" Oct 13 12:47:45 crc kubenswrapper[4678]: I1013 12:47:45.605690 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hkpp7" podStartSLOduration=3.222085428 podStartE2EDuration="4.60567509s" podCreationTimestamp="2025-10-13 12:47:41 +0000 UTC" firstStartedPulling="2025-10-13 12:47:42.532798452 +0000 UTC m=+250.617336376" lastFinishedPulling="2025-10-13 12:47:43.916388154 +0000 UTC m=+252.000926038" observedRunningTime="2025-10-13 12:47:45.602968588 +0000 UTC m=+253.687506462" watchObservedRunningTime="2025-10-13 12:47:45.60567509 +0000 UTC m=+253.690212974" Oct 13 12:47:48 crc kubenswrapper[4678]: I1013 12:47:48.985150 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:48 crc kubenswrapper[4678]: I1013 12:47:48.985512 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.060480 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.222751 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.222847 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.270385 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.650521 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-56qm8" Oct 13 12:47:49 crc kubenswrapper[4678]: I1013 12:47:49.660305 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-brf7p" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.404780 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.405887 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.453823 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.594208 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.594578 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.646126 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:47:51 crc kubenswrapper[4678]: I1013 12:47:51.666840 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b2nmp" Oct 13 12:47:52 crc kubenswrapper[4678]: I1013 12:47:52.677700 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 12:48:55 crc kubenswrapper[4678]: I1013 12:48:55.506563 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:48:55 crc kubenswrapper[4678]: I1013 12:48:55.507806 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:49:25 crc kubenswrapper[4678]: I1013 12:49:25.506802 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:49:25 crc kubenswrapper[4678]: I1013 12:49:25.507529 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:49:55 crc kubenswrapper[4678]: I1013 12:49:55.506297 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:49:55 crc kubenswrapper[4678]: I1013 12:49:55.507024 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:49:55 crc kubenswrapper[4678]: I1013 12:49:55.507126 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:49:55 crc kubenswrapper[4678]: I1013 12:49:55.508017 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 12:49:55 crc kubenswrapper[4678]: I1013 12:49:55.508141 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447" gracePeriod=600 Oct 13 12:49:56 crc kubenswrapper[4678]: I1013 12:49:56.466337 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447" exitCode=0 Oct 13 12:49:56 crc kubenswrapper[4678]: I1013 12:49:56.466443 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447"} Oct 13 12:49:56 crc kubenswrapper[4678]: I1013 12:49:56.467206 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d"} Oct 13 12:49:56 crc kubenswrapper[4678]: I1013 12:49:56.467246 4678 scope.go:117] "RemoveContainer" containerID="118326335aee5d6022d1f9dd6bdc3a67969046c4e5fd3c2b415e326361b3653c" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.075764 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pmb9d"] Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.077380 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.120623 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pmb9d"] Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.274891 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.274945 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-certificates\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.274983 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2cfh\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-kube-api-access-b2cfh\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.275013 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-tls\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.275031 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/958c6e93-fb00-4b8b-89ef-2e3f5e698628-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.275066 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/958c6e93-fb00-4b8b-89ef-2e3f5e698628-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.275085 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-bound-sa-token\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.275103 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-trusted-ca\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.296784 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376028 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2cfh\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-kube-api-access-b2cfh\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376241 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-tls\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376313 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/958c6e93-fb00-4b8b-89ef-2e3f5e698628-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376398 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-bound-sa-token\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376432 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/958c6e93-fb00-4b8b-89ef-2e3f5e698628-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376469 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-trusted-ca\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.376547 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-certificates\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.377559 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/958c6e93-fb00-4b8b-89ef-2e3f5e698628-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.379147 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-trusted-ca\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.380172 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-certificates\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.384616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-registry-tls\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.384912 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/958c6e93-fb00-4b8b-89ef-2e3f5e698628-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.394610 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-bound-sa-token\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.397795 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2cfh\" (UniqueName: \"kubernetes.io/projected/958c6e93-fb00-4b8b-89ef-2e3f5e698628-kube-api-access-b2cfh\") pod \"image-registry-66df7c8f76-pmb9d\" (UID: \"958c6e93-fb00-4b8b-89ef-2e3f5e698628\") " pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.409020 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:40 crc kubenswrapper[4678]: I1013 12:51:40.682899 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pmb9d"] Oct 13 12:51:41 crc kubenswrapper[4678]: I1013 12:51:41.200602 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" event={"ID":"958c6e93-fb00-4b8b-89ef-2e3f5e698628","Type":"ContainerStarted","Data":"00b9b2b86cfc9103b58be70128cb5e36bed548d6d68801e0fb84c8a9aaee0ac2"} Oct 13 12:51:41 crc kubenswrapper[4678]: I1013 12:51:41.201095 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" event={"ID":"958c6e93-fb00-4b8b-89ef-2e3f5e698628","Type":"ContainerStarted","Data":"e9b148c33242e4c380a43042e525cc4eff50ab5b88df53c0246fca3e744ec97c"} Oct 13 12:51:41 crc kubenswrapper[4678]: I1013 12:51:41.201111 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:51:41 crc kubenswrapper[4678]: I1013 12:51:41.229480 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" podStartSLOduration=1.2294612520000001 podStartE2EDuration="1.229461252s" podCreationTimestamp="2025-10-13 12:51:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:51:41.225442607 +0000 UTC m=+489.309980531" watchObservedRunningTime="2025-10-13 12:51:41.229461252 +0000 UTC m=+489.313999136" Oct 13 12:51:55 crc kubenswrapper[4678]: I1013 12:51:55.506552 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:51:55 crc kubenswrapper[4678]: I1013 12:51:55.507462 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:52:00 crc kubenswrapper[4678]: I1013 12:52:00.420313 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-pmb9d" Oct 13 12:52:00 crc kubenswrapper[4678]: I1013 12:52:00.495946 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:52:25 crc kubenswrapper[4678]: I1013 12:52:25.506121 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:52:25 crc kubenswrapper[4678]: I1013 12:52:25.508831 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:52:25 crc kubenswrapper[4678]: I1013 12:52:25.549747 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" podUID="afd9036b-48b5-478a-88e9-d5904ab84534" containerName="registry" containerID="cri-o://8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f" gracePeriod=30 Oct 13 12:52:25 crc kubenswrapper[4678]: I1013 12:52:25.989015 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105195 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105266 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105387 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjdd4\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105434 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105506 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105569 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.105756 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"afd9036b-48b5-478a-88e9-d5904ab84534\" (UID: \"afd9036b-48b5-478a-88e9-d5904ab84534\") " Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.107097 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.108816 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.114448 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.115580 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.117331 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.117625 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4" (OuterVolumeSpecName: "kube-api-access-wjdd4") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "kube-api-access-wjdd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.127109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.137374 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "afd9036b-48b5-478a-88e9-d5904ab84534" (UID: "afd9036b-48b5-478a-88e9-d5904ab84534"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208197 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208265 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208296 4678 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/afd9036b-48b5-478a-88e9-d5904ab84534-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208318 4678 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/afd9036b-48b5-478a-88e9-d5904ab84534-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208337 4678 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208360 4678 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/afd9036b-48b5-478a-88e9-d5904ab84534-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.208382 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjdd4\" (UniqueName: \"kubernetes.io/projected/afd9036b-48b5-478a-88e9-d5904ab84534-kube-api-access-wjdd4\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.508956 4678 generic.go:334] "Generic (PLEG): container finished" podID="afd9036b-48b5-478a-88e9-d5904ab84534" containerID="8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f" exitCode=0 Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.509179 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" event={"ID":"afd9036b-48b5-478a-88e9-d5904ab84534","Type":"ContainerDied","Data":"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f"} Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.510412 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" event={"ID":"afd9036b-48b5-478a-88e9-d5904ab84534","Type":"ContainerDied","Data":"cb30b78a8737dfb89eb5b28d0bf0ca4fe9d05c4e36e04a0666dcb7b13e3b9f1d"} Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.510454 4678 scope.go:117] "RemoveContainer" containerID="8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.509367 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-929f2" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.552433 4678 scope.go:117] "RemoveContainer" containerID="8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.555420 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:52:26 crc kubenswrapper[4678]: E1013 12:52:26.555914 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f\": container with ID starting with 8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f not found: ID does not exist" containerID="8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.555970 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f"} err="failed to get container status \"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f\": rpc error: code = NotFound desc = could not find container \"8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f\": container with ID starting with 8a3fe67e9df5db16b5ee8571ba0a6640321152aca28737dc6c69f2bd07993b6f not found: ID does not exist" Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.564299 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-929f2"] Oct 13 12:52:26 crc kubenswrapper[4678]: I1013 12:52:26.603703 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afd9036b-48b5-478a-88e9-d5904ab84534" path="/var/lib/kubelet/pods/afd9036b-48b5-478a-88e9-d5904ab84534/volumes" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.040687 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-l8c98"] Oct 13 12:52:39 crc kubenswrapper[4678]: E1013 12:52:39.041940 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd9036b-48b5-478a-88e9-d5904ab84534" containerName="registry" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.041974 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd9036b-48b5-478a-88e9-d5904ab84534" containerName="registry" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.042160 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd9036b-48b5-478a-88e9-d5904ab84534" containerName="registry" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.042873 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.049885 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.052803 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-l8c98"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.055733 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-lxm6g" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.055926 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.057636 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vw4gz"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.058645 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vw4gz" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.062391 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-klmk8" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.078496 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vw4gz"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.093788 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pczp4"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.094672 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.097383 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rp9zr" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.102124 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pczp4"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.190212 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvz9c\" (UniqueName: \"kubernetes.io/projected/2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1-kube-api-access-nvz9c\") pod \"cert-manager-cainjector-7f985d654d-l8c98\" (UID: \"2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.190278 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwcl8\" (UniqueName: \"kubernetes.io/projected/e0a73006-9a7f-41d4-a99c-d749fdfad405-kube-api-access-bwcl8\") pod \"cert-manager-5b446d88c5-vw4gz\" (UID: \"e0a73006-9a7f-41d4-a99c-d749fdfad405\") " pod="cert-manager/cert-manager-5b446d88c5-vw4gz" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.291665 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwcl8\" (UniqueName: \"kubernetes.io/projected/e0a73006-9a7f-41d4-a99c-d749fdfad405-kube-api-access-bwcl8\") pod \"cert-manager-5b446d88c5-vw4gz\" (UID: \"e0a73006-9a7f-41d4-a99c-d749fdfad405\") " pod="cert-manager/cert-manager-5b446d88c5-vw4gz" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.291720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29rnc\" (UniqueName: \"kubernetes.io/projected/2d01b76f-171e-4e27-a020-3e43c2405b19-kube-api-access-29rnc\") pod \"cert-manager-webhook-5655c58dd6-pczp4\" (UID: \"2d01b76f-171e-4e27-a020-3e43c2405b19\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.291798 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvz9c\" (UniqueName: \"kubernetes.io/projected/2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1-kube-api-access-nvz9c\") pod \"cert-manager-cainjector-7f985d654d-l8c98\" (UID: \"2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.310587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvz9c\" (UniqueName: \"kubernetes.io/projected/2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1-kube-api-access-nvz9c\") pod \"cert-manager-cainjector-7f985d654d-l8c98\" (UID: \"2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.312811 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwcl8\" (UniqueName: \"kubernetes.io/projected/e0a73006-9a7f-41d4-a99c-d749fdfad405-kube-api-access-bwcl8\") pod \"cert-manager-5b446d88c5-vw4gz\" (UID: \"e0a73006-9a7f-41d4-a99c-d749fdfad405\") " pod="cert-manager/cert-manager-5b446d88c5-vw4gz" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.372346 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.395704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29rnc\" (UniqueName: \"kubernetes.io/projected/2d01b76f-171e-4e27-a020-3e43c2405b19-kube-api-access-29rnc\") pod \"cert-manager-webhook-5655c58dd6-pczp4\" (UID: \"2d01b76f-171e-4e27-a020-3e43c2405b19\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.396259 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vw4gz" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.418469 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29rnc\" (UniqueName: \"kubernetes.io/projected/2d01b76f-171e-4e27-a020-3e43c2405b19-kube-api-access-29rnc\") pod \"cert-manager-webhook-5655c58dd6-pczp4\" (UID: \"2d01b76f-171e-4e27-a020-3e43c2405b19\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.602294 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-l8c98"] Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.612530 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.666321 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vw4gz"] Oct 13 12:52:39 crc kubenswrapper[4678]: W1013 12:52:39.669930 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a73006_9a7f_41d4_a99c_d749fdfad405.slice/crio-ca699ae2d8f6b5aaf15913fb4d9ac81df74249ec395d3c101deaeb6c73eb700d WatchSource:0}: Error finding container ca699ae2d8f6b5aaf15913fb4d9ac81df74249ec395d3c101deaeb6c73eb700d: Status 404 returned error can't find the container with id ca699ae2d8f6b5aaf15913fb4d9ac81df74249ec395d3c101deaeb6c73eb700d Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.711489 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:39 crc kubenswrapper[4678]: I1013 12:52:39.936784 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-pczp4"] Oct 13 12:52:39 crc kubenswrapper[4678]: W1013 12:52:39.942147 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d01b76f_171e_4e27_a020_3e43c2405b19.slice/crio-adfe2bd7e6a2b1f5c76d6579168c04b1c036bb454efc612b1be49c3babd7689b WatchSource:0}: Error finding container adfe2bd7e6a2b1f5c76d6579168c04b1c036bb454efc612b1be49c3babd7689b: Status 404 returned error can't find the container with id adfe2bd7e6a2b1f5c76d6579168c04b1c036bb454efc612b1be49c3babd7689b Oct 13 12:52:40 crc kubenswrapper[4678]: I1013 12:52:40.616949 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" event={"ID":"2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1","Type":"ContainerStarted","Data":"dc272ac933d9b273cbe70a458d0d84cb8009138d963cbdb138c3881827f26f47"} Oct 13 12:52:40 crc kubenswrapper[4678]: I1013 12:52:40.622377 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vw4gz" event={"ID":"e0a73006-9a7f-41d4-a99c-d749fdfad405","Type":"ContainerStarted","Data":"ca699ae2d8f6b5aaf15913fb4d9ac81df74249ec395d3c101deaeb6c73eb700d"} Oct 13 12:52:40 crc kubenswrapper[4678]: I1013 12:52:40.626283 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" event={"ID":"2d01b76f-171e-4e27-a020-3e43c2405b19","Type":"ContainerStarted","Data":"adfe2bd7e6a2b1f5c76d6579168c04b1c036bb454efc612b1be49c3babd7689b"} Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.666873 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" event={"ID":"2d01b76f-171e-4e27-a020-3e43c2405b19","Type":"ContainerStarted","Data":"0005d36839b1a9dc51eacfd7490ee481e500529c77ff52ad35df6edb03f4d012"} Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.667550 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.671119 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" event={"ID":"2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1","Type":"ContainerStarted","Data":"398d649537880cd20619464104d0ba7612607fdabf27eca0a6fb5d2bb85e1dab"} Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.674039 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vw4gz" event={"ID":"e0a73006-9a7f-41d4-a99c-d749fdfad405","Type":"ContainerStarted","Data":"aa463bc611bcdc91002f6165df79a581643f4459a1782b49b9f22f8e49e09015"} Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.695861 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" podStartSLOduration=2.104350714 podStartE2EDuration="5.69583095s" podCreationTimestamp="2025-10-13 12:52:39 +0000 UTC" firstStartedPulling="2025-10-13 12:52:39.944021137 +0000 UTC m=+548.028559021" lastFinishedPulling="2025-10-13 12:52:43.535501363 +0000 UTC m=+551.620039257" observedRunningTime="2025-10-13 12:52:44.695733398 +0000 UTC m=+552.780271322" watchObservedRunningTime="2025-10-13 12:52:44.69583095 +0000 UTC m=+552.780368864" Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.725385 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-l8c98" podStartSLOduration=2.099664191 podStartE2EDuration="5.725350717s" podCreationTimestamp="2025-10-13 12:52:39 +0000 UTC" firstStartedPulling="2025-10-13 12:52:39.61230261 +0000 UTC m=+547.696840484" lastFinishedPulling="2025-10-13 12:52:43.237989126 +0000 UTC m=+551.322527010" observedRunningTime="2025-10-13 12:52:44.718478536 +0000 UTC m=+552.803016450" watchObservedRunningTime="2025-10-13 12:52:44.725350717 +0000 UTC m=+552.809888641" Oct 13 12:52:44 crc kubenswrapper[4678]: I1013 12:52:44.743936 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vw4gz" podStartSLOduration=1.9625344729999998 podStartE2EDuration="5.743890005s" podCreationTimestamp="2025-10-13 12:52:39 +0000 UTC" firstStartedPulling="2025-10-13 12:52:39.672503194 +0000 UTC m=+547.757041078" lastFinishedPulling="2025-10-13 12:52:43.453858676 +0000 UTC m=+551.538396610" observedRunningTime="2025-10-13 12:52:44.743211077 +0000 UTC m=+552.827749031" watchObservedRunningTime="2025-10-13 12:52:44.743890005 +0000 UTC m=+552.828427919" Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.653762 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hsx89"] Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.654811 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-controller" containerID="cri-o://29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.654934 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="northd" containerID="cri-o://adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.654997 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-node" containerID="cri-o://38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.655012 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="sbdb" containerID="cri-o://12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.654975 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.655092 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-acl-logging" containerID="cri-o://1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.655201 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="nbdb" containerID="cri-o://4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: W1013 12:52:49.687326 4678 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331.scope/cpu.max": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331.scope/cpu.max: no such device Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.699740 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" containerID="cri-o://07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" gracePeriod=30 Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.720115 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-pczp4" Oct 13 12:52:49 crc kubenswrapper[4678]: E1013 12:52:49.797616 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-conmon-1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21d5191d_621a_4416_b635_e3459b3cfe8d.slice/crio-94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-conmon-29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod43e78486_f760_44e0_a85e_f3868869ae5b.slice/crio-conmon-38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21d5191d_621a_4416_b635_e3459b3cfe8d.slice/crio-conmon-94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb.scope\": RecentStats: unable to find data in memory cache]" Oct 13 12:52:49 crc kubenswrapper[4678]: I1013 12:52:49.997915 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/3.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.001307 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovn-acl-logging/0.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.001793 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovn-controller/0.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.002255 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.072934 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-l8bwp"] Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073374 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073410 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073431 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="sbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073445 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="sbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073467 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-node" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073482 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-node" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073499 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kubecfg-setup" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073511 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kubecfg-setup" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073527 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073540 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073559 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073571 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073589 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="nbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073601 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="nbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073620 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073632 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073648 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="northd" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073660 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="northd" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073683 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-acl-logging" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073697 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-acl-logging" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073715 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073728 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.073751 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073763 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073971 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.073990 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074004 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074022 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="northd" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074043 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074098 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovn-acl-logging" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074117 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="sbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074135 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-ovn-metrics" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074153 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="nbdb" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074169 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="kube-rbac-proxy-node" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.074339 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074354 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074548 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.074578 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" containerName="ovnkube-controller" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.077122 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173177 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173343 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173447 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173718 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173833 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.173970 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash" (OuterVolumeSpecName: "host-slash") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174149 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174244 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174255 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174300 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174354 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174370 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174409 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174410 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log" (OuterVolumeSpecName: "node-log") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174465 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174560 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174688 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174730 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174773 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174810 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174874 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174692 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174888 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174807 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket" (OuterVolumeSpecName: "log-socket") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174843 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174898 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174746 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174908 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.174984 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175031 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4tf8\" (UniqueName: \"kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175128 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175205 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175479 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch\") pod \"43e78486-f760-44e0-a85e-f3868869ae5b\" (UID: \"43e78486-f760-44e0-a85e-f3868869ae5b\") " Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175618 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175683 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-var-lib-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175703 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175735 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-netd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175775 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-ovn\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175810 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-etc-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175840 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-log-socket\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175849 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175874 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-slash\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.175959 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x4fq\" (UniqueName: \"kubernetes.io/projected/2a0eeaf7-e99b-47ae-a987-db2f02af6797-kube-api-access-5x4fq\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176074 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-node-log\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176120 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-config\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-systemd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176193 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-kubelet\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176245 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovn-node-metrics-cert\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176273 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-env-overrides\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176315 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-systemd-units\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176512 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-script-lib\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176615 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-netns\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176660 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-bin\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176716 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176744 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176842 4678 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176859 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176870 4678 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176882 4678 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-node-log\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176895 4678 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176907 4678 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176917 4678 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176927 4678 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-log-socket\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176936 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176946 4678 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176955 4678 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176964 4678 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176975 4678 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176984 4678 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.176993 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/43e78486-f760-44e0-a85e-f3868869ae5b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.177004 4678 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.177015 4678 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-host-slash\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.211964 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.213172 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.213249 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8" (OuterVolumeSpecName: "kube-api-access-q4tf8") pod "43e78486-f760-44e0-a85e-f3868869ae5b" (UID: "43e78486-f760-44e0-a85e-f3868869ae5b"). InnerVolumeSpecName "kube-api-access-q4tf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278399 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278519 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278536 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278573 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-var-lib-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278601 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-ovn-kubernetes\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-netd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278632 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-var-lib-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-ovn\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278660 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-netd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278684 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-etc-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278694 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-ovn\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278716 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-log-socket\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278751 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-slash\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278784 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x4fq\" (UniqueName: \"kubernetes.io/projected/2a0eeaf7-e99b-47ae-a987-db2f02af6797-kube-api-access-5x4fq\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278796 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-log-socket\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278803 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-etc-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278865 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-slash\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278894 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-node-log\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278927 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-config\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278964 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-systemd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.278995 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-kubelet\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279025 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovn-node-metrics-cert\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279039 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-node-log\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279094 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-kubelet\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279060 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279071 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-systemd\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279137 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-env-overrides\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279080 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-run-openvswitch\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279176 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-systemd-units\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279216 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-systemd-units\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279240 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-script-lib\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-netns\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279446 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-bin\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279542 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-run-netns\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279606 4678 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/43e78486-f760-44e0-a85e-f3868869ae5b-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279644 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4tf8\" (UniqueName: \"kubernetes.io/projected/43e78486-f760-44e0-a85e-f3868869ae5b-kube-api-access-q4tf8\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279672 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/43e78486-f760-44e0-a85e-f3868869ae5b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279714 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2a0eeaf7-e99b-47ae-a987-db2f02af6797-host-cni-bin\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.279885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-env-overrides\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.280457 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-script-lib\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.280732 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovnkube-config\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.283964 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2a0eeaf7-e99b-47ae-a987-db2f02af6797-ovn-node-metrics-cert\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.311510 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x4fq\" (UniqueName: \"kubernetes.io/projected/2a0eeaf7-e99b-47ae-a987-db2f02af6797-kube-api-access-5x4fq\") pod \"ovnkube-node-l8bwp\" (UID: \"2a0eeaf7-e99b-47ae-a987-db2f02af6797\") " pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.397856 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.722801 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovnkube-controller/3.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.725447 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovn-acl-logging/0.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726239 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hsx89_43e78486-f760-44e0-a85e-f3868869ae5b/ovn-controller/0.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726679 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726698 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726707 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726714 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726721 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726728 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726734 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" exitCode=143 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726748 4678 generic.go:334] "Generic (PLEG): container finished" podID="43e78486-f760-44e0-a85e-f3868869ae5b" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" exitCode=143 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726788 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726839 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726849 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726859 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726877 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726886 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726891 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726897 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726902 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726907 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726913 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726918 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726923 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726939 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726945 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726950 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726955 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726960 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726965 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726970 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726975 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726980 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726985 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726991 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.726998 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727004 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727009 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727014 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727019 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727024 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727028 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727098 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727104 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727109 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" event={"ID":"43e78486-f760-44e0-a85e-f3868869ae5b","Type":"ContainerDied","Data":"40bad1607225e7f28937bdc9d263bc6e82cd37a6d4e5bb243b3834c213745b20"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727128 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727134 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727140 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727146 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727159 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727164 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727170 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727175 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727181 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727187 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727200 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.727351 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hsx89" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.733258 4678 generic.go:334] "Generic (PLEG): container finished" podID="2a0eeaf7-e99b-47ae-a987-db2f02af6797" containerID="a9f00d979db28dfbcab48a019aee0b55828e210e99ef0df6f4b376663a2c5739" exitCode=0 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.733364 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerDied","Data":"a9f00d979db28dfbcab48a019aee0b55828e210e99ef0df6f4b376663a2c5739"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.733455 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"b1f8097cb1be47c482d9958931b2a46edff4720e1f29ebafb34209c5fd29dc4e"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.737161 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/2.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.737818 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/1.log" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.737876 4678 generic.go:334] "Generic (PLEG): container finished" podID="21d5191d-621a-4416-b635-e3459b3cfe8d" containerID="94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb" exitCode=2 Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.737914 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerDied","Data":"94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.737949 4678 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c"} Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.738623 4678 scope.go:117] "RemoveContainer" containerID="94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb" Oct 13 12:52:50 crc kubenswrapper[4678]: E1013 12:52:50.738925 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nlfg7_openshift-multus(21d5191d-621a-4416-b635-e3459b3cfe8d)\"" pod="openshift-multus/multus-nlfg7" podUID="21d5191d-621a-4416-b635-e3459b3cfe8d" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.754947 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hsx89"] Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.757338 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hsx89"] Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.757415 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.846946 4678 scope.go:117] "RemoveContainer" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.868451 4678 scope.go:117] "RemoveContainer" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.886744 4678 scope.go:117] "RemoveContainer" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.910114 4678 scope.go:117] "RemoveContainer" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.926214 4678 scope.go:117] "RemoveContainer" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.942744 4678 scope.go:117] "RemoveContainer" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.968659 4678 scope.go:117] "RemoveContainer" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:50 crc kubenswrapper[4678]: I1013 12:52:50.989689 4678 scope.go:117] "RemoveContainer" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.011140 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.011654 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.011714 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} err="failed to get container status \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.011749 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.012081 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": container with ID starting with b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa not found: ID does not exist" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012120 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} err="failed to get container status \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": rpc error: code = NotFound desc = could not find container \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": container with ID starting with b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012148 4678 scope.go:117] "RemoveContainer" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.012465 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": container with ID starting with 12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728 not found: ID does not exist" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012532 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} err="failed to get container status \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": rpc error: code = NotFound desc = could not find container \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": container with ID starting with 12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012571 4678 scope.go:117] "RemoveContainer" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.012914 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": container with ID starting with 4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b not found: ID does not exist" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012959 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} err="failed to get container status \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": rpc error: code = NotFound desc = could not find container \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": container with ID starting with 4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.012986 4678 scope.go:117] "RemoveContainer" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.013343 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": container with ID starting with adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721 not found: ID does not exist" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.013379 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} err="failed to get container status \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": rpc error: code = NotFound desc = could not find container \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": container with ID starting with adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.013395 4678 scope.go:117] "RemoveContainer" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.013683 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": container with ID starting with a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7 not found: ID does not exist" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.013724 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} err="failed to get container status \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": rpc error: code = NotFound desc = could not find container \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": container with ID starting with a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.013754 4678 scope.go:117] "RemoveContainer" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.014021 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": container with ID starting with 38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29 not found: ID does not exist" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014064 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} err="failed to get container status \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": rpc error: code = NotFound desc = could not find container \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": container with ID starting with 38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014080 4678 scope.go:117] "RemoveContainer" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.014362 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": container with ID starting with 1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9 not found: ID does not exist" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014405 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} err="failed to get container status \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": rpc error: code = NotFound desc = could not find container \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": container with ID starting with 1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014431 4678 scope.go:117] "RemoveContainer" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.014711 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": container with ID starting with 29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331 not found: ID does not exist" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014733 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} err="failed to get container status \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": rpc error: code = NotFound desc = could not find container \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": container with ID starting with 29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.014746 4678 scope.go:117] "RemoveContainer" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: E1013 12:52:51.015189 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": container with ID starting with 449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7 not found: ID does not exist" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015227 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} err="failed to get container status \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": rpc error: code = NotFound desc = could not find container \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": container with ID starting with 449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015255 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015594 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} err="failed to get container status \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015638 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015918 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} err="failed to get container status \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": rpc error: code = NotFound desc = could not find container \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": container with ID starting with b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.015941 4678 scope.go:117] "RemoveContainer" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016204 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} err="failed to get container status \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": rpc error: code = NotFound desc = could not find container \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": container with ID starting with 12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016231 4678 scope.go:117] "RemoveContainer" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016536 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} err="failed to get container status \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": rpc error: code = NotFound desc = could not find container \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": container with ID starting with 4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016584 4678 scope.go:117] "RemoveContainer" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016890 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} err="failed to get container status \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": rpc error: code = NotFound desc = could not find container \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": container with ID starting with adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.016926 4678 scope.go:117] "RemoveContainer" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.017214 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} err="failed to get container status \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": rpc error: code = NotFound desc = could not find container \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": container with ID starting with a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.017252 4678 scope.go:117] "RemoveContainer" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.017560 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} err="failed to get container status \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": rpc error: code = NotFound desc = could not find container \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": container with ID starting with 38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.017601 4678 scope.go:117] "RemoveContainer" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018163 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} err="failed to get container status \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": rpc error: code = NotFound desc = could not find container \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": container with ID starting with 1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018193 4678 scope.go:117] "RemoveContainer" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018498 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} err="failed to get container status \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": rpc error: code = NotFound desc = could not find container \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": container with ID starting with 29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018536 4678 scope.go:117] "RemoveContainer" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018814 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} err="failed to get container status \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": rpc error: code = NotFound desc = could not find container \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": container with ID starting with 449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.018858 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019219 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} err="failed to get container status \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019272 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019564 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} err="failed to get container status \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": rpc error: code = NotFound desc = could not find container \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": container with ID starting with b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019601 4678 scope.go:117] "RemoveContainer" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019867 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} err="failed to get container status \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": rpc error: code = NotFound desc = could not find container \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": container with ID starting with 12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.019907 4678 scope.go:117] "RemoveContainer" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.020191 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} err="failed to get container status \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": rpc error: code = NotFound desc = could not find container \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": container with ID starting with 4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.020228 4678 scope.go:117] "RemoveContainer" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.020894 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} err="failed to get container status \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": rpc error: code = NotFound desc = could not find container \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": container with ID starting with adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.020929 4678 scope.go:117] "RemoveContainer" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.021226 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} err="failed to get container status \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": rpc error: code = NotFound desc = could not find container \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": container with ID starting with a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.021265 4678 scope.go:117] "RemoveContainer" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.021495 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} err="failed to get container status \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": rpc error: code = NotFound desc = could not find container \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": container with ID starting with 38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.021532 4678 scope.go:117] "RemoveContainer" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.022504 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} err="failed to get container status \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": rpc error: code = NotFound desc = could not find container \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": container with ID starting with 1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.022541 4678 scope.go:117] "RemoveContainer" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.023417 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} err="failed to get container status \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": rpc error: code = NotFound desc = could not find container \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": container with ID starting with 29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.023456 4678 scope.go:117] "RemoveContainer" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.023771 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} err="failed to get container status \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": rpc error: code = NotFound desc = could not find container \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": container with ID starting with 449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.023808 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.024147 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} err="failed to get container status \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.024178 4678 scope.go:117] "RemoveContainer" containerID="b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.026109 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa"} err="failed to get container status \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": rpc error: code = NotFound desc = could not find container \"b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa\": container with ID starting with b116559e414389a1b72f2784e961560c1a4eb929158ca340f15a0a7693f845aa not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.026141 4678 scope.go:117] "RemoveContainer" containerID="12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.026499 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728"} err="failed to get container status \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": rpc error: code = NotFound desc = could not find container \"12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728\": container with ID starting with 12f6617612aafdd3dcc8fd4022ba5bc722c4057b0f18b92cfbe334332fb03728 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.026535 4678 scope.go:117] "RemoveContainer" containerID="4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027194 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b"} err="failed to get container status \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": rpc error: code = NotFound desc = could not find container \"4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b\": container with ID starting with 4a8a42d7a8004a62b201f306fdbd5fafbb762139e16ba83e268bea1971e5325b not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027229 4678 scope.go:117] "RemoveContainer" containerID="adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027535 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721"} err="failed to get container status \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": rpc error: code = NotFound desc = could not find container \"adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721\": container with ID starting with adeec10b1f71373382656f3cf84a5462588b9d2fde69560344e74272d74b8721 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027567 4678 scope.go:117] "RemoveContainer" containerID="a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027940 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7"} err="failed to get container status \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": rpc error: code = NotFound desc = could not find container \"a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7\": container with ID starting with a26f12eb4681efff84e152f4175fa75bba747844b1b29268580f745738f5c9b7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.027969 4678 scope.go:117] "RemoveContainer" containerID="38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.028463 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29"} err="failed to get container status \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": rpc error: code = NotFound desc = could not find container \"38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29\": container with ID starting with 38a4739622c95d45a4898b485350c87d976e150d5f999eae6bbc04c1380eed29 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.028491 4678 scope.go:117] "RemoveContainer" containerID="1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.028856 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9"} err="failed to get container status \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": rpc error: code = NotFound desc = could not find container \"1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9\": container with ID starting with 1b7342cb9483dc9893b885bdc1efa9173ff088106c9f503096b68c71f30a3de9 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.028877 4678 scope.go:117] "RemoveContainer" containerID="29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.029415 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331"} err="failed to get container status \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": rpc error: code = NotFound desc = could not find container \"29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331\": container with ID starting with 29494db97e4ef8a22a0c86f3094169a57c784065f3fcf3aceb962b1cea807331 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.029507 4678 scope.go:117] "RemoveContainer" containerID="449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.029942 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7"} err="failed to get container status \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": rpc error: code = NotFound desc = could not find container \"449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7\": container with ID starting with 449466c51ac39540410ab2d9a6f2d712d9235902155c7e4de13b7917510f49c7 not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.029980 4678 scope.go:117] "RemoveContainer" containerID="07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.030289 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb"} err="failed to get container status \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": rpc error: code = NotFound desc = could not find container \"07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb\": container with ID starting with 07ef9d2e7503bcb4e753057982df2a056b59efd02ef0d32df786c6972c71b9eb not found: ID does not exist" Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750198 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"4dfcb674642c6eeb8dff00c33518c27c4dfaf68d55851488efe79f725456df27"} Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750577 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"c07dcab902532590bfca10a7b5af23bf1be199bf6a512aa4cf67a89fc0041068"} Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750591 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"24800b80897115be7c8056c851b368c0759189fda5f46e770806cbcb8dd42ac3"} Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750603 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"5d502e46acf47745148e89aa4ecac671dfa3dac6390f64406896958d7eebed4d"} Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750617 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"23e8e0723af717b39073c4b6facc7c32ece625fa4b38bf237bd1c12f4f3490d4"} Oct 13 12:52:51 crc kubenswrapper[4678]: I1013 12:52:51.750628 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"ca38d723ff0de32c41b29811ed9616191b9c1449af8d3544f8f91b018e53b1e3"} Oct 13 12:52:52 crc kubenswrapper[4678]: I1013 12:52:52.607219 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43e78486-f760-44e0-a85e-f3868869ae5b" path="/var/lib/kubelet/pods/43e78486-f760-44e0-a85e-f3868869ae5b/volumes" Oct 13 12:52:54 crc kubenswrapper[4678]: I1013 12:52:54.778221 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"306f721bcc340d74773798cffe1e8bf0487dc88cf607ada79d23afd94e64b6c5"} Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.506308 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.506638 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.506679 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.507242 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.507303 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d" gracePeriod=600 Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.787104 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d" exitCode=0 Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.787148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d"} Oct 13 12:52:55 crc kubenswrapper[4678]: I1013 12:52:55.787188 4678 scope.go:117] "RemoveContainer" containerID="5a9c9eca909184e92cf25cd97fb7f83b5fbc53daf19d24b23ec4bca022a31447" Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.794452 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" event={"ID":"2a0eeaf7-e99b-47ae-a987-db2f02af6797","Type":"ContainerStarted","Data":"49df7e2d0259a67e3786084f4f4cb3fea8983038d6a69dbf39fd719826b07aa0"} Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.795032 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.795066 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.798655 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002"} Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.819364 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" podStartSLOduration=6.819345917 podStartE2EDuration="6.819345917s" podCreationTimestamp="2025-10-13 12:52:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:52:56.816711738 +0000 UTC m=+564.901249622" watchObservedRunningTime="2025-10-13 12:52:56.819345917 +0000 UTC m=+564.903883801" Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.828483 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:56 crc kubenswrapper[4678]: I1013 12:52:56.828560 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:52:57 crc kubenswrapper[4678]: I1013 12:52:57.803159 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:53:05 crc kubenswrapper[4678]: I1013 12:53:05.592739 4678 scope.go:117] "RemoveContainer" containerID="94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb" Oct 13 12:53:05 crc kubenswrapper[4678]: E1013 12:53:05.593909 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-nlfg7_openshift-multus(21d5191d-621a-4416-b635-e3459b3cfe8d)\"" pod="openshift-multus/multus-nlfg7" podUID="21d5191d-621a-4416-b635-e3459b3cfe8d" Oct 13 12:53:18 crc kubenswrapper[4678]: I1013 12:53:18.592924 4678 scope.go:117] "RemoveContainer" containerID="94d47bf22dc1b3d6ee605c3a169b936ee1e284d66126ae17b5e12db209ddffdb" Oct 13 12:53:18 crc kubenswrapper[4678]: I1013 12:53:18.941725 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/2.log" Oct 13 12:53:18 crc kubenswrapper[4678]: I1013 12:53:18.942888 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/1.log" Oct 13 12:53:18 crc kubenswrapper[4678]: I1013 12:53:18.942970 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nlfg7" event={"ID":"21d5191d-621a-4416-b635-e3459b3cfe8d","Type":"ContainerStarted","Data":"b96c14ab249756e2d06c3a08183602c0c3d80d6a28c4566c1243700dde951b25"} Oct 13 12:53:20 crc kubenswrapper[4678]: I1013 12:53:20.434314 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-l8bwp" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.034986 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh"] Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.037863 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.042196 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.046169 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh"] Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.143943 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.144035 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8kgk\" (UniqueName: \"kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.144242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.245639 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8kgk\" (UniqueName: \"kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.245730 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.245777 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.246275 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.246647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.269556 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8kgk\" (UniqueName: \"kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.354995 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:30 crc kubenswrapper[4678]: I1013 12:53:30.619652 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh"] Oct 13 12:53:30 crc kubenswrapper[4678]: W1013 12:53:30.630268 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fa7ade0_a26e_45ea_ab14_017c299f5582.slice/crio-01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c WatchSource:0}: Error finding container 01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c: Status 404 returned error can't find the container with id 01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c Oct 13 12:53:31 crc kubenswrapper[4678]: I1013 12:53:31.019281 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerStarted","Data":"469cda4d7d453b9d13d87949c808076df6969d8454b0fd430d61ad8f53bb2fe2"} Oct 13 12:53:31 crc kubenswrapper[4678]: I1013 12:53:31.019342 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerStarted","Data":"01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c"} Oct 13 12:53:32 crc kubenswrapper[4678]: I1013 12:53:32.030666 4678 generic.go:334] "Generic (PLEG): container finished" podID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerID="469cda4d7d453b9d13d87949c808076df6969d8454b0fd430d61ad8f53bb2fe2" exitCode=0 Oct 13 12:53:32 crc kubenswrapper[4678]: I1013 12:53:32.030743 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerDied","Data":"469cda4d7d453b9d13d87949c808076df6969d8454b0fd430d61ad8f53bb2fe2"} Oct 13 12:53:32 crc kubenswrapper[4678]: I1013 12:53:32.818003 4678 scope.go:117] "RemoveContainer" containerID="1046999945e9dbdbe6026a39a717661e1666a2cc28d46606d8c3fa8534a6b15c" Oct 13 12:53:33 crc kubenswrapper[4678]: I1013 12:53:33.039314 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nlfg7_21d5191d-621a-4416-b635-e3459b3cfe8d/kube-multus/2.log" Oct 13 12:53:35 crc kubenswrapper[4678]: I1013 12:53:35.054854 4678 generic.go:334] "Generic (PLEG): container finished" podID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerID="ae584da695939ea4e37027e71f0d8ad4ee54373a575d405dacb07d7fee97636f" exitCode=0 Oct 13 12:53:35 crc kubenswrapper[4678]: I1013 12:53:35.054926 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerDied","Data":"ae584da695939ea4e37027e71f0d8ad4ee54373a575d405dacb07d7fee97636f"} Oct 13 12:53:36 crc kubenswrapper[4678]: I1013 12:53:36.066147 4678 generic.go:334] "Generic (PLEG): container finished" podID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerID="4444daa12461cc2f3009417d12b189a3b3fbdad3f04c93ddf1e1b02de4501d4c" exitCode=0 Oct 13 12:53:36 crc kubenswrapper[4678]: I1013 12:53:36.066194 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerDied","Data":"4444daa12461cc2f3009417d12b189a3b3fbdad3f04c93ddf1e1b02de4501d4c"} Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.430583 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.583279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8kgk\" (UniqueName: \"kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk\") pod \"9fa7ade0-a26e-45ea-ab14-017c299f5582\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.583365 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle\") pod \"9fa7ade0-a26e-45ea-ab14-017c299f5582\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.583421 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util\") pod \"9fa7ade0-a26e-45ea-ab14-017c299f5582\" (UID: \"9fa7ade0-a26e-45ea-ab14-017c299f5582\") " Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.585156 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle" (OuterVolumeSpecName: "bundle") pod "9fa7ade0-a26e-45ea-ab14-017c299f5582" (UID: "9fa7ade0-a26e-45ea-ab14-017c299f5582"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.594511 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk" (OuterVolumeSpecName: "kube-api-access-z8kgk") pod "9fa7ade0-a26e-45ea-ab14-017c299f5582" (UID: "9fa7ade0-a26e-45ea-ab14-017c299f5582"). InnerVolumeSpecName "kube-api-access-z8kgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.605823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util" (OuterVolumeSpecName: "util") pod "9fa7ade0-a26e-45ea-ab14-017c299f5582" (UID: "9fa7ade0-a26e-45ea-ab14-017c299f5582"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.685444 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8kgk\" (UniqueName: \"kubernetes.io/projected/9fa7ade0-a26e-45ea-ab14-017c299f5582-kube-api-access-z8kgk\") on node \"crc\" DevicePath \"\"" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.685511 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:53:37 crc kubenswrapper[4678]: I1013 12:53:37.685531 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fa7ade0-a26e-45ea-ab14-017c299f5582-util\") on node \"crc\" DevicePath \"\"" Oct 13 12:53:38 crc kubenswrapper[4678]: I1013 12:53:38.084155 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" event={"ID":"9fa7ade0-a26e-45ea-ab14-017c299f5582","Type":"ContainerDied","Data":"01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c"} Oct 13 12:53:38 crc kubenswrapper[4678]: I1013 12:53:38.084196 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01779ac7ef513f446e7e9037897c9c3cde8dd8b37f43cf989acb7bfd4b34833c" Oct 13 12:53:38 crc kubenswrapper[4678]: I1013 12:53:38.084227 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.649872 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hk95b"] Oct 13 12:53:41 crc kubenswrapper[4678]: E1013 12:53:41.650442 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="pull" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.650457 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="pull" Oct 13 12:53:41 crc kubenswrapper[4678]: E1013 12:53:41.650481 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="extract" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.650489 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="extract" Oct 13 12:53:41 crc kubenswrapper[4678]: E1013 12:53:41.650502 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="util" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.650510 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="util" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.650646 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fa7ade0-a26e-45ea-ab14-017c299f5582" containerName="extract" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.651097 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.653304 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.654816 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4sl4c" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.654941 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.670822 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hk95b"] Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.743098 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf52m\" (UniqueName: \"kubernetes.io/projected/0a00ef2f-1663-499a-97a7-1b14e7d7f3c1-kube-api-access-bf52m\") pod \"nmstate-operator-858ddd8f98-hk95b\" (UID: \"0a00ef2f-1663-499a-97a7-1b14e7d7f3c1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.844636 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf52m\" (UniqueName: \"kubernetes.io/projected/0a00ef2f-1663-499a-97a7-1b14e7d7f3c1-kube-api-access-bf52m\") pod \"nmstate-operator-858ddd8f98-hk95b\" (UID: \"0a00ef2f-1663-499a-97a7-1b14e7d7f3c1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" Oct 13 12:53:41 crc kubenswrapper[4678]: I1013 12:53:41.862989 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf52m\" (UniqueName: \"kubernetes.io/projected/0a00ef2f-1663-499a-97a7-1b14e7d7f3c1-kube-api-access-bf52m\") pod \"nmstate-operator-858ddd8f98-hk95b\" (UID: \"0a00ef2f-1663-499a-97a7-1b14e7d7f3c1\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" Oct 13 12:53:42 crc kubenswrapper[4678]: I1013 12:53:42.004699 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" Oct 13 12:53:42 crc kubenswrapper[4678]: W1013 12:53:42.270650 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a00ef2f_1663_499a_97a7_1b14e7d7f3c1.slice/crio-a9fbfcc6943958add05e764ffe64cc456279c0fd39df80aef243c1f72f31bbbd WatchSource:0}: Error finding container a9fbfcc6943958add05e764ffe64cc456279c0fd39df80aef243c1f72f31bbbd: Status 404 returned error can't find the container with id a9fbfcc6943958add05e764ffe64cc456279c0fd39df80aef243c1f72f31bbbd Oct 13 12:53:42 crc kubenswrapper[4678]: I1013 12:53:42.270739 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hk95b"] Oct 13 12:53:43 crc kubenswrapper[4678]: I1013 12:53:43.117378 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" event={"ID":"0a00ef2f-1663-499a-97a7-1b14e7d7f3c1","Type":"ContainerStarted","Data":"a9fbfcc6943958add05e764ffe64cc456279c0fd39df80aef243c1f72f31bbbd"} Oct 13 12:53:45 crc kubenswrapper[4678]: I1013 12:53:45.138063 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" event={"ID":"0a00ef2f-1663-499a-97a7-1b14e7d7f3c1","Type":"ContainerStarted","Data":"a9ab8258882eb288c9e7e2b943f4c661ae39c6818547f0a6667bca9c00779b7f"} Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.385527 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hk95b" podStartSLOduration=7.202938776 podStartE2EDuration="9.385499482s" podCreationTimestamp="2025-10-13 12:53:41 +0000 UTC" firstStartedPulling="2025-10-13 12:53:42.273000416 +0000 UTC m=+610.357538310" lastFinishedPulling="2025-10-13 12:53:44.455561122 +0000 UTC m=+612.540099016" observedRunningTime="2025-10-13 12:53:45.165645448 +0000 UTC m=+613.250183372" watchObservedRunningTime="2025-10-13 12:53:50.385499482 +0000 UTC m=+618.470037376" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.386295 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.387470 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.390073 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.390980 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.391671 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-r4ctm" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.391725 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.395531 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.445426 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.459283 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-rtdjd"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.461547 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475519 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475581 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s82xw\" (UniqueName: \"kubernetes.io/projected/ae486b1a-e0b6-45ff-91f2-6bb967b88436-kube-api-access-s82xw\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475604 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-dbus-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-ovs-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475713 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dm74\" (UniqueName: \"kubernetes.io/projected/00b6c406-a244-403d-9ca0-2406de2da1f5-kube-api-access-8dm74\") pod \"nmstate-metrics-fdff9cb8d-nws2x\" (UID: \"00b6c406-a244-403d-9ca0-2406de2da1f5\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475741 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-nmstate-lock\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.475776 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb4vq\" (UniqueName: \"kubernetes.io/projected/5439ee39-b0d6-43fc-92aa-f5240c460afc-kube-api-access-vb4vq\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.550390 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.551362 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.554907 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7xwmp" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.554936 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.557679 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.560515 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576825 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576871 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576896 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e3884e53-e229-44e5-8d8c-42fcc654f657-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576915 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s82xw\" (UniqueName: \"kubernetes.io/projected/ae486b1a-e0b6-45ff-91f2-6bb967b88436-kube-api-access-s82xw\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576932 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-dbus-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: E1013 12:53:50.576996 4678 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577035 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-ovs-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.576997 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-ovs-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: E1013 12:53:50.577084 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair podName:ae486b1a-e0b6-45ff-91f2-6bb967b88436 nodeName:}" failed. No retries permitted until 2025-10-13 12:53:51.077044858 +0000 UTC m=+619.161582752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair") pod "nmstate-webhook-6cdbc54649-tbt2v" (UID: "ae486b1a-e0b6-45ff-91f2-6bb967b88436") : secret "openshift-nmstate-webhook" not found Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wmvp\" (UniqueName: \"kubernetes.io/projected/e3884e53-e229-44e5-8d8c-42fcc654f657-kube-api-access-8wmvp\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577216 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-dbus-socket\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dm74\" (UniqueName: \"kubernetes.io/projected/00b6c406-a244-403d-9ca0-2406de2da1f5-kube-api-access-8dm74\") pod \"nmstate-metrics-fdff9cb8d-nws2x\" (UID: \"00b6c406-a244-403d-9ca0-2406de2da1f5\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577270 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-nmstate-lock\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577314 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb4vq\" (UniqueName: \"kubernetes.io/projected/5439ee39-b0d6-43fc-92aa-f5240c460afc-kube-api-access-vb4vq\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.577385 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5439ee39-b0d6-43fc-92aa-f5240c460afc-nmstate-lock\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.594596 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dm74\" (UniqueName: \"kubernetes.io/projected/00b6c406-a244-403d-9ca0-2406de2da1f5-kube-api-access-8dm74\") pod \"nmstate-metrics-fdff9cb8d-nws2x\" (UID: \"00b6c406-a244-403d-9ca0-2406de2da1f5\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.594639 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s82xw\" (UniqueName: \"kubernetes.io/projected/ae486b1a-e0b6-45ff-91f2-6bb967b88436-kube-api-access-s82xw\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.595029 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb4vq\" (UniqueName: \"kubernetes.io/projected/5439ee39-b0d6-43fc-92aa-f5240c460afc-kube-api-access-vb4vq\") pod \"nmstate-handler-rtdjd\" (UID: \"5439ee39-b0d6-43fc-92aa-f5240c460afc\") " pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.678122 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.678168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e3884e53-e229-44e5-8d8c-42fcc654f657-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.678193 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wmvp\" (UniqueName: \"kubernetes.io/projected/e3884e53-e229-44e5-8d8c-42fcc654f657-kube-api-access-8wmvp\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: E1013 12:53:50.678314 4678 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 13 12:53:50 crc kubenswrapper[4678]: E1013 12:53:50.678391 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert podName:e3884e53-e229-44e5-8d8c-42fcc654f657 nodeName:}" failed. No retries permitted until 2025-10-13 12:53:51.178368544 +0000 UTC m=+619.262906428 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-c5csj" (UID: "e3884e53-e229-44e5-8d8c-42fcc654f657") : secret "plugin-serving-cert" not found Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.679254 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/e3884e53-e229-44e5-8d8c-42fcc654f657-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.694322 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wmvp\" (UniqueName: \"kubernetes.io/projected/e3884e53-e229-44e5-8d8c-42fcc654f657-kube-api-access-8wmvp\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.739546 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-d5d6b78c7-dfcpk"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.740646 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.748116 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-d5d6b78c7-dfcpk"] Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.778890 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-oauth-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.778957 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-service-ca\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.778983 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-trusted-ca-bundle\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.779092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.779112 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-oauth-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.779135 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-console-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.779312 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpjjw\" (UniqueName: \"kubernetes.io/projected/e1cc872f-e198-49c8-ac54-620457cba412-kube-api-access-qpjjw\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.793673 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.810807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.880993 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-oauth-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.881276 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-service-ca\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.881296 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-trusted-ca-bundle\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.881341 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.881357 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-oauth-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.881377 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-console-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.882127 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-service-ca\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.882199 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-oauth-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.882231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpjjw\" (UniqueName: \"kubernetes.io/projected/e1cc872f-e198-49c8-ac54-620457cba412-kube-api-access-qpjjw\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.882705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-console-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.882940 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cc872f-e198-49c8-ac54-620457cba412-trusted-ca-bundle\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.885608 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-serving-cert\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.886534 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e1cc872f-e198-49c8-ac54-620457cba412-console-oauth-config\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:50 crc kubenswrapper[4678]: I1013 12:53:50.901003 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpjjw\" (UniqueName: \"kubernetes.io/projected/e1cc872f-e198-49c8-ac54-620457cba412-kube-api-access-qpjjw\") pod \"console-d5d6b78c7-dfcpk\" (UID: \"e1cc872f-e198-49c8-ac54-620457cba412\") " pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.059534 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.085419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.093282 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ae486b1a-e0b6-45ff-91f2-6bb967b88436-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-tbt2v\" (UID: \"ae486b1a-e0b6-45ff-91f2-6bb967b88436\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.177713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rtdjd" event={"ID":"5439ee39-b0d6-43fc-92aa-f5240c460afc","Type":"ContainerStarted","Data":"683771c7c276ab893a49d1d1c5ff62b31faace6c0cd040beb8f9421cfcfe13ef"} Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.186481 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.193594 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3884e53-e229-44e5-8d8c-42fcc654f657-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-c5csj\" (UID: \"e3884e53-e229-44e5-8d8c-42fcc654f657\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.204651 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x"] Oct 13 12:53:51 crc kubenswrapper[4678]: W1013 12:53:51.214362 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00b6c406_a244_403d_9ca0_2406de2da1f5.slice/crio-16251c16c46366b924648ceda4a0c1ac6e6d39f65dc9a794f9098bb0deb1fc04 WatchSource:0}: Error finding container 16251c16c46366b924648ceda4a0c1ac6e6d39f65dc9a794f9098bb0deb1fc04: Status 404 returned error can't find the container with id 16251c16c46366b924648ceda4a0c1ac6e6d39f65dc9a794f9098bb0deb1fc04 Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.378430 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.464006 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.535786 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-d5d6b78c7-dfcpk"] Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.691644 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj"] Oct 13 12:53:51 crc kubenswrapper[4678]: W1013 12:53:51.699278 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3884e53_e229_44e5_8d8c_42fcc654f657.slice/crio-4552a17fd4d4859aa610af61fa8792cdc3f9bc54825f46291b46ca1524c986a1 WatchSource:0}: Error finding container 4552a17fd4d4859aa610af61fa8792cdc3f9bc54825f46291b46ca1524c986a1: Status 404 returned error can't find the container with id 4552a17fd4d4859aa610af61fa8792cdc3f9bc54825f46291b46ca1524c986a1 Oct 13 12:53:51 crc kubenswrapper[4678]: I1013 12:53:51.836998 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v"] Oct 13 12:53:51 crc kubenswrapper[4678]: W1013 12:53:51.846266 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae486b1a_e0b6_45ff_91f2_6bb967b88436.slice/crio-8776a64f973f7b99d465dae9a6ddfcd782dc0e810a61b9f226aefa29d90b9342 WatchSource:0}: Error finding container 8776a64f973f7b99d465dae9a6ddfcd782dc0e810a61b9f226aefa29d90b9342: Status 404 returned error can't find the container with id 8776a64f973f7b99d465dae9a6ddfcd782dc0e810a61b9f226aefa29d90b9342 Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.184436 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" event={"ID":"e3884e53-e229-44e5-8d8c-42fcc654f657","Type":"ContainerStarted","Data":"4552a17fd4d4859aa610af61fa8792cdc3f9bc54825f46291b46ca1524c986a1"} Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.185470 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" event={"ID":"ae486b1a-e0b6-45ff-91f2-6bb967b88436","Type":"ContainerStarted","Data":"8776a64f973f7b99d465dae9a6ddfcd782dc0e810a61b9f226aefa29d90b9342"} Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.186722 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d5d6b78c7-dfcpk" event={"ID":"e1cc872f-e198-49c8-ac54-620457cba412","Type":"ContainerStarted","Data":"7046aaf9a4c6488e6b1810ea3573dab2db1cd27452d58f08a8f35bca8e76c671"} Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.186753 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d5d6b78c7-dfcpk" event={"ID":"e1cc872f-e198-49c8-ac54-620457cba412","Type":"ContainerStarted","Data":"0ec6f0fba3e619d795ef6e4fe4c97a7184767317802bb60c5e3d44ae231356a4"} Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.187882 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" event={"ID":"00b6c406-a244-403d-9ca0-2406de2da1f5","Type":"ContainerStarted","Data":"16251c16c46366b924648ceda4a0c1ac6e6d39f65dc9a794f9098bb0deb1fc04"} Oct 13 12:53:52 crc kubenswrapper[4678]: I1013 12:53:52.211326 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-d5d6b78c7-dfcpk" podStartSLOduration=2.211302757 podStartE2EDuration="2.211302757s" podCreationTimestamp="2025-10-13 12:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:53:52.201465733 +0000 UTC m=+620.286003627" watchObservedRunningTime="2025-10-13 12:53:52.211302757 +0000 UTC m=+620.295840681" Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.205560 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" event={"ID":"e3884e53-e229-44e5-8d8c-42fcc654f657","Type":"ContainerStarted","Data":"431bbdad408505e2b07cee70f848f721844bb8bc5d1b6954c923a7c30223fcb5"} Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.207381 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" event={"ID":"ae486b1a-e0b6-45ff-91f2-6bb967b88436","Type":"ContainerStarted","Data":"26f28f04ea69f99171ea40da2d3206e6242051b505f140e954b9b55c746d57d6"} Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.207519 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.208881 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-rtdjd" event={"ID":"5439ee39-b0d6-43fc-92aa-f5240c460afc","Type":"ContainerStarted","Data":"4e21ec8567cbfad75f854994762ea3b741664306f04736265ed7f4470fbd92dd"} Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.209010 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.210091 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" event={"ID":"00b6c406-a244-403d-9ca0-2406de2da1f5","Type":"ContainerStarted","Data":"4863cbdd530dd447437955e77ca11c4a06b90c349478f4f884ead12496277a8b"} Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.222851 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-c5csj" podStartSLOduration=2.680105272 podStartE2EDuration="5.222834179s" podCreationTimestamp="2025-10-13 12:53:50 +0000 UTC" firstStartedPulling="2025-10-13 12:53:51.702372015 +0000 UTC m=+619.786909899" lastFinishedPulling="2025-10-13 12:53:54.245100902 +0000 UTC m=+622.329638806" observedRunningTime="2025-10-13 12:53:55.220378105 +0000 UTC m=+623.304915989" watchObservedRunningTime="2025-10-13 12:53:55.222834179 +0000 UTC m=+623.307372063" Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.246251 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-rtdjd" podStartSLOduration=1.8245872109999999 podStartE2EDuration="5.246235683s" podCreationTimestamp="2025-10-13 12:53:50 +0000 UTC" firstStartedPulling="2025-10-13 12:53:50.83272985 +0000 UTC m=+618.917267734" lastFinishedPulling="2025-10-13 12:53:54.254378272 +0000 UTC m=+622.338916206" observedRunningTime="2025-10-13 12:53:55.243137683 +0000 UTC m=+623.327675577" watchObservedRunningTime="2025-10-13 12:53:55.246235683 +0000 UTC m=+623.330773567" Oct 13 12:53:55 crc kubenswrapper[4678]: I1013 12:53:55.276960 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" podStartSLOduration=2.845805721 podStartE2EDuration="5.276937026s" podCreationTimestamp="2025-10-13 12:53:50 +0000 UTC" firstStartedPulling="2025-10-13 12:53:51.853435886 +0000 UTC m=+619.937973770" lastFinishedPulling="2025-10-13 12:53:54.284567181 +0000 UTC m=+622.369105075" observedRunningTime="2025-10-13 12:53:55.272597984 +0000 UTC m=+623.357135868" watchObservedRunningTime="2025-10-13 12:53:55.276937026 +0000 UTC m=+623.361474940" Oct 13 12:54:00 crc kubenswrapper[4678]: I1013 12:54:00.829717 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-rtdjd" Oct 13 12:54:01 crc kubenswrapper[4678]: I1013 12:54:01.059975 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:54:01 crc kubenswrapper[4678]: I1013 12:54:01.060591 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:54:01 crc kubenswrapper[4678]: I1013 12:54:01.065724 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:54:01 crc kubenswrapper[4678]: I1013 12:54:01.257852 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-d5d6b78c7-dfcpk" Oct 13 12:54:01 crc kubenswrapper[4678]: I1013 12:54:01.332513 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:54:02 crc kubenswrapper[4678]: I1013 12:54:02.260772 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" event={"ID":"00b6c406-a244-403d-9ca0-2406de2da1f5","Type":"ContainerStarted","Data":"64b268ed4653dc3b01b078b267cfd9229a4a89a378e4c5daa6be004de471d638"} Oct 13 12:54:02 crc kubenswrapper[4678]: I1013 12:54:02.287140 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-nws2x" podStartSLOduration=1.624600947 podStartE2EDuration="12.287118689s" podCreationTimestamp="2025-10-13 12:53:50 +0000 UTC" firstStartedPulling="2025-10-13 12:53:51.217085214 +0000 UTC m=+619.301623138" lastFinishedPulling="2025-10-13 12:54:01.879602956 +0000 UTC m=+629.964140880" observedRunningTime="2025-10-13 12:54:02.281379801 +0000 UTC m=+630.365917725" watchObservedRunningTime="2025-10-13 12:54:02.287118689 +0000 UTC m=+630.371656613" Oct 13 12:54:11 crc kubenswrapper[4678]: I1013 12:54:11.388829 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-tbt2v" Oct 13 12:54:26 crc kubenswrapper[4678]: I1013 12:54:26.431657 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-qq6wd" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerName="console" containerID="cri-o://ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a" gracePeriod=15 Oct 13 12:54:26 crc kubenswrapper[4678]: I1013 12:54:26.831101 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qq6wd_bfb45e1a-27cb-4199-9a3d-657962b29d32/console/0.log" Oct 13 12:54:26 crc kubenswrapper[4678]: I1013 12:54:26.831427 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015589 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015664 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnbsr\" (UniqueName: \"kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015781 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015829 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015860 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.015900 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca\") pod \"bfb45e1a-27cb-4199-9a3d-657962b29d32\" (UID: \"bfb45e1a-27cb-4199-9a3d-657962b29d32\") " Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.017253 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.017347 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca" (OuterVolumeSpecName: "service-ca") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.018262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config" (OuterVolumeSpecName: "console-config") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.018348 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.022266 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.022617 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.023913 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr" (OuterVolumeSpecName: "kube-api-access-gnbsr") pod "bfb45e1a-27cb-4199-9a3d-657962b29d32" (UID: "bfb45e1a-27cb-4199-9a3d-657962b29d32"). InnerVolumeSpecName "kube-api-access-gnbsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117896 4678 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117927 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnbsr\" (UniqueName: \"kubernetes.io/projected/bfb45e1a-27cb-4199-9a3d-657962b29d32-kube-api-access-gnbsr\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117939 4678 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117949 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117958 4678 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117966 4678 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-console-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.117975 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bfb45e1a-27cb-4199-9a3d-657962b29d32-service-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453662 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qq6wd_bfb45e1a-27cb-4199-9a3d-657962b29d32/console/0.log" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453726 4678 generic.go:334] "Generic (PLEG): container finished" podID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerID="ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a" exitCode=2 Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453762 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq6wd" event={"ID":"bfb45e1a-27cb-4199-9a3d-657962b29d32","Type":"ContainerDied","Data":"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a"} Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453793 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq6wd" event={"ID":"bfb45e1a-27cb-4199-9a3d-657962b29d32","Type":"ContainerDied","Data":"01333799daaba45dd3f234aea023684422ee5b17f250c1f198a98789489dbaa6"} Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453813 4678 scope.go:117] "RemoveContainer" containerID="ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.453831 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq6wd" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.492880 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.494798 4678 scope.go:117] "RemoveContainer" containerID="ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a" Oct 13 12:54:27 crc kubenswrapper[4678]: E1013 12:54:27.495229 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a\": container with ID starting with ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a not found: ID does not exist" containerID="ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.495280 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a"} err="failed to get container status \"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a\": rpc error: code = NotFound desc = could not find container \"ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a\": container with ID starting with ba5b4000e51032089ac8a91f837ec81f27452123c49dddcbeae4879cb7c3ce0a not found: ID does not exist" Oct 13 12:54:27 crc kubenswrapper[4678]: I1013 12:54:27.497245 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-qq6wd"] Oct 13 12:54:28 crc kubenswrapper[4678]: I1013 12:54:28.601810 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" path="/var/lib/kubelet/pods/bfb45e1a-27cb-4199-9a3d-657962b29d32/volumes" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.649415 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4"] Oct 13 12:54:29 crc kubenswrapper[4678]: E1013 12:54:29.650224 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerName="console" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.650248 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerName="console" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.650436 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb45e1a-27cb-4199-9a3d-657962b29d32" containerName="console" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.651754 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.654673 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.658428 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4"] Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.753889 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44x4x\" (UniqueName: \"kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.754020 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.754283 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.856615 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44x4x\" (UniqueName: \"kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.856748 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.856809 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.857726 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.858221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.889514 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44x4x\" (UniqueName: \"kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:29 crc kubenswrapper[4678]: I1013 12:54:29.971303 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:30 crc kubenswrapper[4678]: I1013 12:54:30.275517 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4"] Oct 13 12:54:30 crc kubenswrapper[4678]: I1013 12:54:30.477114 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" event={"ID":"d05a6818-fca2-4ab9-8d9a-4aceb001c74c","Type":"ContainerStarted","Data":"6f32a1295c6828aacccacbde0a7b5b241bcb3b56945dc394542713014ccb1a24"} Oct 13 12:54:31 crc kubenswrapper[4678]: I1013 12:54:31.488471 4678 generic.go:334] "Generic (PLEG): container finished" podID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerID="79a709f62283fb5b6a418ee36c405b8f307f6fc6fc981dddf250ffd27eedab74" exitCode=0 Oct 13 12:54:31 crc kubenswrapper[4678]: I1013 12:54:31.488769 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" event={"ID":"d05a6818-fca2-4ab9-8d9a-4aceb001c74c","Type":"ContainerDied","Data":"79a709f62283fb5b6a418ee36c405b8f307f6fc6fc981dddf250ffd27eedab74"} Oct 13 12:54:33 crc kubenswrapper[4678]: I1013 12:54:33.507639 4678 generic.go:334] "Generic (PLEG): container finished" podID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerID="5660f1ff6b62d3ad40518499461d69e20b731a865f269ea58509d0cfb6202fcd" exitCode=0 Oct 13 12:54:33 crc kubenswrapper[4678]: I1013 12:54:33.507757 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" event={"ID":"d05a6818-fca2-4ab9-8d9a-4aceb001c74c","Type":"ContainerDied","Data":"5660f1ff6b62d3ad40518499461d69e20b731a865f269ea58509d0cfb6202fcd"} Oct 13 12:54:34 crc kubenswrapper[4678]: I1013 12:54:34.517821 4678 generic.go:334] "Generic (PLEG): container finished" podID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerID="c34e5832a650fb5e8565f16f6822d12749cc49edeb76e87accd2acf76d8bd6a0" exitCode=0 Oct 13 12:54:34 crc kubenswrapper[4678]: I1013 12:54:34.517892 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" event={"ID":"d05a6818-fca2-4ab9-8d9a-4aceb001c74c","Type":"ContainerDied","Data":"c34e5832a650fb5e8565f16f6822d12749cc49edeb76e87accd2acf76d8bd6a0"} Oct 13 12:54:35 crc kubenswrapper[4678]: I1013 12:54:35.854294 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.052426 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44x4x\" (UniqueName: \"kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x\") pod \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.052556 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util\") pod \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.052645 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle\") pod \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\" (UID: \"d05a6818-fca2-4ab9-8d9a-4aceb001c74c\") " Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.053865 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle" (OuterVolumeSpecName: "bundle") pod "d05a6818-fca2-4ab9-8d9a-4aceb001c74c" (UID: "d05a6818-fca2-4ab9-8d9a-4aceb001c74c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.059442 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x" (OuterVolumeSpecName: "kube-api-access-44x4x") pod "d05a6818-fca2-4ab9-8d9a-4aceb001c74c" (UID: "d05a6818-fca2-4ab9-8d9a-4aceb001c74c"). InnerVolumeSpecName "kube-api-access-44x4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.145046 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util" (OuterVolumeSpecName: "util") pod "d05a6818-fca2-4ab9-8d9a-4aceb001c74c" (UID: "d05a6818-fca2-4ab9-8d9a-4aceb001c74c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.155069 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.155119 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44x4x\" (UniqueName: \"kubernetes.io/projected/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-kube-api-access-44x4x\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.155139 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d05a6818-fca2-4ab9-8d9a-4aceb001c74c-util\") on node \"crc\" DevicePath \"\"" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.535835 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" event={"ID":"d05a6818-fca2-4ab9-8d9a-4aceb001c74c","Type":"ContainerDied","Data":"6f32a1295c6828aacccacbde0a7b5b241bcb3b56945dc394542713014ccb1a24"} Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.535884 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f32a1295c6828aacccacbde0a7b5b241bcb3b56945dc394542713014ccb1a24" Oct 13 12:54:36 crc kubenswrapper[4678]: I1013 12:54:36.535905 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.175730 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4"] Oct 13 12:54:45 crc kubenswrapper[4678]: E1013 12:54:45.176659 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="util" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.176678 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="util" Oct 13 12:54:45 crc kubenswrapper[4678]: E1013 12:54:45.176696 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="extract" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.176704 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="extract" Oct 13 12:54:45 crc kubenswrapper[4678]: E1013 12:54:45.176722 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="pull" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.176729 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="pull" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.176856 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05a6818-fca2-4ab9-8d9a-4aceb001c74c" containerName="extract" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.177549 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.180993 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.181041 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.183090 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.183357 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.183487 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fsrpq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.195686 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4"] Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.264664 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z72ct\" (UniqueName: \"kubernetes.io/projected/51bd00ff-af4f-4405-9f25-20b4d22e7020-kube-api-access-z72ct\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.264750 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-apiservice-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.264777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-webhook-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.365192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-apiservice-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.365235 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-webhook-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.365277 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z72ct\" (UniqueName: \"kubernetes.io/projected/51bd00ff-af4f-4405-9f25-20b4d22e7020-kube-api-access-z72ct\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.371110 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-webhook-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.387774 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z72ct\" (UniqueName: \"kubernetes.io/projected/51bd00ff-af4f-4405-9f25-20b4d22e7020-kube-api-access-z72ct\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.388023 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51bd00ff-af4f-4405-9f25-20b4d22e7020-apiservice-cert\") pod \"metallb-operator-controller-manager-797dff698d-tbgv4\" (UID: \"51bd00ff-af4f-4405-9f25-20b4d22e7020\") " pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.416029 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq"] Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.416651 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.423394 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.423602 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-2g2mt" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.424082 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.432012 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq"] Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.503633 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.567673 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9czk7\" (UniqueName: \"kubernetes.io/projected/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-kube-api-access-9czk7\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.567738 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-webhook-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.567765 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-apiservice-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.669650 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9czk7\" (UniqueName: \"kubernetes.io/projected/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-kube-api-access-9czk7\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.669698 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-webhook-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.669722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-apiservice-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.676041 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-apiservice-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.679693 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-webhook-cert\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.689747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9czk7\" (UniqueName: \"kubernetes.io/projected/0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5-kube-api-access-9czk7\") pod \"metallb-operator-webhook-server-6545fd778d-8xqwq\" (UID: \"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5\") " pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.770478 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.824835 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4"] Oct 13 12:54:45 crc kubenswrapper[4678]: W1013 12:54:45.836651 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51bd00ff_af4f_4405_9f25_20b4d22e7020.slice/crio-7452a6195cb67906365a5e7f5a5e8d19720420fbf0694d48e6df49a80a8c1439 WatchSource:0}: Error finding container 7452a6195cb67906365a5e7f5a5e8d19720420fbf0694d48e6df49a80a8c1439: Status 404 returned error can't find the container with id 7452a6195cb67906365a5e7f5a5e8d19720420fbf0694d48e6df49a80a8c1439 Oct 13 12:54:45 crc kubenswrapper[4678]: I1013 12:54:45.980868 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq"] Oct 13 12:54:45 crc kubenswrapper[4678]: W1013 12:54:45.999352 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b593c83_b8a6_4cf2_8dbe_245b3d3abfc5.slice/crio-d854589ab33a39a27396beea4eb097f798a410df195470c78cfa9d30f74841ee WatchSource:0}: Error finding container d854589ab33a39a27396beea4eb097f798a410df195470c78cfa9d30f74841ee: Status 404 returned error can't find the container with id d854589ab33a39a27396beea4eb097f798a410df195470c78cfa9d30f74841ee Oct 13 12:54:46 crc kubenswrapper[4678]: I1013 12:54:46.607094 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" event={"ID":"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5","Type":"ContainerStarted","Data":"d854589ab33a39a27396beea4eb097f798a410df195470c78cfa9d30f74841ee"} Oct 13 12:54:46 crc kubenswrapper[4678]: I1013 12:54:46.609117 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" event={"ID":"51bd00ff-af4f-4405-9f25-20b4d22e7020","Type":"ContainerStarted","Data":"7452a6195cb67906365a5e7f5a5e8d19720420fbf0694d48e6df49a80a8c1439"} Oct 13 12:54:55 crc kubenswrapper[4678]: I1013 12:54:55.505791 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:54:55 crc kubenswrapper[4678]: I1013 12:54:55.506813 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:54:57 crc kubenswrapper[4678]: I1013 12:54:57.667734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" event={"ID":"0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5","Type":"ContainerStarted","Data":"b3efb3be98a58632e3fde8e72cc9c7e533df72b922bda1d7e68f8f4e114b8de3"} Oct 13 12:54:57 crc kubenswrapper[4678]: I1013 12:54:57.668472 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:54:57 crc kubenswrapper[4678]: I1013 12:54:57.687743 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" podStartSLOduration=2.013941209 podStartE2EDuration="12.68772262s" podCreationTimestamp="2025-10-13 12:54:45 +0000 UTC" firstStartedPulling="2025-10-13 12:54:46.002549444 +0000 UTC m=+674.087087318" lastFinishedPulling="2025-10-13 12:54:56.676330845 +0000 UTC m=+684.760868729" observedRunningTime="2025-10-13 12:54:57.685882243 +0000 UTC m=+685.770420167" watchObservedRunningTime="2025-10-13 12:54:57.68772262 +0000 UTC m=+685.772260504" Oct 13 12:55:12 crc kubenswrapper[4678]: I1013 12:55:12.761285 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" event={"ID":"51bd00ff-af4f-4405-9f25-20b4d22e7020","Type":"ContainerStarted","Data":"a4dc7ea26743f248fb05782605cd1e1e76aa6b617173affaff6a30eae3554686"} Oct 13 12:55:12 crc kubenswrapper[4678]: I1013 12:55:12.762935 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:55:12 crc kubenswrapper[4678]: I1013 12:55:12.793699 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" podStartSLOduration=1.823093102 podStartE2EDuration="27.793680839s" podCreationTimestamp="2025-10-13 12:54:45 +0000 UTC" firstStartedPulling="2025-10-13 12:54:45.838868718 +0000 UTC m=+673.923406632" lastFinishedPulling="2025-10-13 12:55:11.809456475 +0000 UTC m=+699.893994369" observedRunningTime="2025-10-13 12:55:12.791100542 +0000 UTC m=+700.875638436" watchObservedRunningTime="2025-10-13 12:55:12.793680839 +0000 UTC m=+700.878218733" Oct 13 12:55:15 crc kubenswrapper[4678]: I1013 12:55:15.776117 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6545fd778d-8xqwq" Oct 13 12:55:25 crc kubenswrapper[4678]: I1013 12:55:25.505872 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:55:25 crc kubenswrapper[4678]: I1013 12:55:25.506317 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:55:45 crc kubenswrapper[4678]: I1013 12:55:45.506861 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-797dff698d-tbgv4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.362284 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xp775"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.363123 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.365243 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hrvlj" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.365590 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.378889 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hdzf4"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.381893 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.383512 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.384170 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.387882 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xp775"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.456917 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mbsh9"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.457921 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.462413 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.462453 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.462806 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5gjgt" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.462942 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.474569 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-x9tg5"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.475529 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.481146 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.489040 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-x9tg5"] Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542202 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-sockets\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542235 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-conf\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542262 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542583 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-startup\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542856 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics-certs\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542882 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phc4q\" (UniqueName: \"kubernetes.io/projected/ead738f8-1d8c-491d-9c14-8cf4b535badf-kube-api-access-phc4q\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542901 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjq4k\" (UniqueName: \"kubernetes.io/projected/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-kube-api-access-bjq4k\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.542928 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-reloader\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644089 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644138 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-startup\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644183 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-cert\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644207 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hnvw\" (UniqueName: \"kubernetes.io/projected/f0f04693-55fb-40f0-8f97-c91d12a02509-kube-api-access-9hnvw\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644227 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics-certs\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phc4q\" (UniqueName: \"kubernetes.io/projected/ead738f8-1d8c-491d-9c14-8cf4b535badf-kube-api-access-phc4q\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjq4k\" (UniqueName: \"kubernetes.io/projected/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-kube-api-access-bjq4k\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644391 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-reloader\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644440 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-metrics-certs\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644481 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644529 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f0f04693-55fb-40f0-8f97-c91d12a02509-metallb-excludel2\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-sockets\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644639 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-conf\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644667 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtgbd\" (UniqueName: \"kubernetes.io/projected/ae7e1bbe-3d07-40db-991b-797c25efd443-kube-api-access-qtgbd\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644692 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.644989 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.645107 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-sockets\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.645313 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-startup\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.645427 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-frr-conf\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.645662 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ead738f8-1d8c-491d-9c14-8cf4b535badf-reloader\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.651245 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-cert\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.662225 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ead738f8-1d8c-491d-9c14-8cf4b535badf-metrics-certs\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.664805 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjq4k\" (UniqueName: \"kubernetes.io/projected/61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac-kube-api-access-bjq4k\") pod \"frr-k8s-webhook-server-64bf5d555-xp775\" (UID: \"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.678819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phc4q\" (UniqueName: \"kubernetes.io/projected/ead738f8-1d8c-491d-9c14-8cf4b535badf-kube-api-access-phc4q\") pod \"frr-k8s-hdzf4\" (UID: \"ead738f8-1d8c-491d-9c14-8cf4b535badf\") " pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.691577 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.704870 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtgbd\" (UniqueName: \"kubernetes.io/projected/ae7e1bbe-3d07-40db-991b-797c25efd443-kube-api-access-qtgbd\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745319 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745345 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745376 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-cert\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745394 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hnvw\" (UniqueName: \"kubernetes.io/projected/f0f04693-55fb-40f0-8f97-c91d12a02509-kube-api-access-9hnvw\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745472 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-metrics-certs\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.745503 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f0f04693-55fb-40f0-8f97-c91d12a02509-metallb-excludel2\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: E1013 12:55:46.745689 4678 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 13 12:55:46 crc kubenswrapper[4678]: E1013 12:55:46.745797 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist podName:f0f04693-55fb-40f0-8f97-c91d12a02509 nodeName:}" failed. No retries permitted until 2025-10-13 12:55:47.245779874 +0000 UTC m=+735.330317758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist") pod "speaker-mbsh9" (UID: "f0f04693-55fb-40f0-8f97-c91d12a02509") : secret "metallb-memberlist" not found Oct 13 12:55:46 crc kubenswrapper[4678]: E1013 12:55:46.745992 4678 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 13 12:55:46 crc kubenswrapper[4678]: E1013 12:55:46.746091 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs podName:ae7e1bbe-3d07-40db-991b-797c25efd443 nodeName:}" failed. No retries permitted until 2025-10-13 12:55:47.246083452 +0000 UTC m=+735.330621326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs") pod "controller-68d546b9d8-x9tg5" (UID: "ae7e1bbe-3d07-40db-991b-797c25efd443") : secret "controller-certs-secret" not found Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.746443 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f0f04693-55fb-40f0-8f97-c91d12a02509-metallb-excludel2\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.748860 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.755159 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-metrics-certs\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.767834 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtgbd\" (UniqueName: \"kubernetes.io/projected/ae7e1bbe-3d07-40db-991b-797c25efd443-kube-api-access-qtgbd\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.768040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-cert\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.771664 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hnvw\" (UniqueName: \"kubernetes.io/projected/f0f04693-55fb-40f0-8f97-c91d12a02509-kube-api-access-9hnvw\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.878566 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-xp775"] Oct 13 12:55:46 crc kubenswrapper[4678]: W1013 12:55:46.883906 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61e5f2b7_9ff7_4c33_a4ff_bbc07fbea5ac.slice/crio-1d911856779ac6ad19b643e15fc0c302762014c8dedd9e6567034e30cc45f06e WatchSource:0}: Error finding container 1d911856779ac6ad19b643e15fc0c302762014c8dedd9e6567034e30cc45f06e: Status 404 returned error can't find the container with id 1d911856779ac6ad19b643e15fc0c302762014c8dedd9e6567034e30cc45f06e Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.998957 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" event={"ID":"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac","Type":"ContainerStarted","Data":"1d911856779ac6ad19b643e15fc0c302762014c8dedd9e6567034e30cc45f06e"} Oct 13 12:55:46 crc kubenswrapper[4678]: I1013 12:55:46.999865 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"96c4342ee0cdbd691bff38359eda3956be08e76dc57361fc97e4362efd1f059a"} Oct 13 12:55:47 crc kubenswrapper[4678]: I1013 12:55:47.252434 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:47 crc kubenswrapper[4678]: E1013 12:55:47.252618 4678 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 13 12:55:47 crc kubenswrapper[4678]: I1013 12:55:47.252678 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:47 crc kubenswrapper[4678]: E1013 12:55:47.252692 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist podName:f0f04693-55fb-40f0-8f97-c91d12a02509 nodeName:}" failed. No retries permitted until 2025-10-13 12:55:48.252671622 +0000 UTC m=+736.337209516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist") pod "speaker-mbsh9" (UID: "f0f04693-55fb-40f0-8f97-c91d12a02509") : secret "metallb-memberlist" not found Oct 13 12:55:47 crc kubenswrapper[4678]: I1013 12:55:47.258937 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e1bbe-3d07-40db-991b-797c25efd443-metrics-certs\") pod \"controller-68d546b9d8-x9tg5\" (UID: \"ae7e1bbe-3d07-40db-991b-797c25efd443\") " pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:47 crc kubenswrapper[4678]: I1013 12:55:47.395140 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:47 crc kubenswrapper[4678]: I1013 12:55:47.710804 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-x9tg5"] Oct 13 12:55:48 crc kubenswrapper[4678]: I1013 12:55:48.007441 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-x9tg5" event={"ID":"ae7e1bbe-3d07-40db-991b-797c25efd443","Type":"ContainerStarted","Data":"7b02dca95c327aed1480eedca537ae2be5b188fe55d6e4859bf7079d2d392da7"} Oct 13 12:55:48 crc kubenswrapper[4678]: I1013 12:55:48.007503 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-x9tg5" event={"ID":"ae7e1bbe-3d07-40db-991b-797c25efd443","Type":"ContainerStarted","Data":"b6a14f6a6f7868b898832476463a2fcca3a402846bfa9bf443337f6c5a155944"} Oct 13 12:55:48 crc kubenswrapper[4678]: I1013 12:55:48.275016 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:48 crc kubenswrapper[4678]: I1013 12:55:48.288082 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f0f04693-55fb-40f0-8f97-c91d12a02509-memberlist\") pod \"speaker-mbsh9\" (UID: \"f0f04693-55fb-40f0-8f97-c91d12a02509\") " pod="metallb-system/speaker-mbsh9" Oct 13 12:55:48 crc kubenswrapper[4678]: I1013 12:55:48.573508 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mbsh9" Oct 13 12:55:49 crc kubenswrapper[4678]: I1013 12:55:49.013243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-x9tg5" event={"ID":"ae7e1bbe-3d07-40db-991b-797c25efd443","Type":"ContainerStarted","Data":"26042ec91910a18b86a416ab4a980b66ea02186389a7d62f15c4354562226c3f"} Oct 13 12:55:49 crc kubenswrapper[4678]: I1013 12:55:49.013855 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:49 crc kubenswrapper[4678]: I1013 12:55:49.017045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mbsh9" event={"ID":"f0f04693-55fb-40f0-8f97-c91d12a02509","Type":"ContainerStarted","Data":"96b561fdfdd7b4d583155efb3d400f6ee207e98deb7c96ea61285494c8fd46ff"} Oct 13 12:55:49 crc kubenswrapper[4678]: I1013 12:55:49.017089 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mbsh9" event={"ID":"f0f04693-55fb-40f0-8f97-c91d12a02509","Type":"ContainerStarted","Data":"85be31e6d8841086c0e4f08056b04ebcafb4669172d0e2ca86f3b14c8e50f0d3"} Oct 13 12:55:49 crc kubenswrapper[4678]: I1013 12:55:49.030477 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-x9tg5" podStartSLOduration=3.030461252 podStartE2EDuration="3.030461252s" podCreationTimestamp="2025-10-13 12:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:55:49.028214413 +0000 UTC m=+737.112752297" watchObservedRunningTime="2025-10-13 12:55:49.030461252 +0000 UTC m=+737.114999136" Oct 13 12:55:50 crc kubenswrapper[4678]: I1013 12:55:50.024471 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mbsh9" event={"ID":"f0f04693-55fb-40f0-8f97-c91d12a02509","Type":"ContainerStarted","Data":"9e3a7f8cc0e8af25ceaa0625c3dbf614c57a63b831e12d8d9f94bb973ce5b95e"} Oct 13 12:55:50 crc kubenswrapper[4678]: I1013 12:55:50.024515 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mbsh9" Oct 13 12:55:52 crc kubenswrapper[4678]: I1013 12:55:52.607990 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mbsh9" podStartSLOduration=6.60797448 podStartE2EDuration="6.60797448s" podCreationTimestamp="2025-10-13 12:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:55:50.040454309 +0000 UTC m=+738.124992193" watchObservedRunningTime="2025-10-13 12:55:52.60797448 +0000 UTC m=+740.692512364" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.069171 4678 generic.go:334] "Generic (PLEG): container finished" podID="ead738f8-1d8c-491d-9c14-8cf4b535badf" containerID="941af1fd78d57f8925fbac4bc50032aae715915644720757eddb245ff3ca9027" exitCode=0 Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.069247 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerDied","Data":"941af1fd78d57f8925fbac4bc50032aae715915644720757eddb245ff3ca9027"} Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.071019 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" event={"ID":"61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac","Type":"ContainerStarted","Data":"f04a78b6235f794535a51f71d58d37eb1e752ef0e7e3e7b337eb5983788e8b22"} Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.072175 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.141840 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" podStartSLOduration=2.026593466 podStartE2EDuration="9.141815224s" podCreationTimestamp="2025-10-13 12:55:46 +0000 UTC" firstStartedPulling="2025-10-13 12:55:46.886011184 +0000 UTC m=+734.970549068" lastFinishedPulling="2025-10-13 12:55:54.001232942 +0000 UTC m=+742.085770826" observedRunningTime="2025-10-13 12:55:55.133033113 +0000 UTC m=+743.217571027" watchObservedRunningTime="2025-10-13 12:55:55.141815224 +0000 UTC m=+743.226353148" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.506125 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.506186 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.506234 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.506948 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 12:55:55 crc kubenswrapper[4678]: I1013 12:55:55.507031 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002" gracePeriod=600 Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.078352 4678 generic.go:334] "Generic (PLEG): container finished" podID="ead738f8-1d8c-491d-9c14-8cf4b535badf" containerID="f81bcba504de02f46efd9f213fc1b9bf320f6ed872185ada8ac96c9e5d2ca74e" exitCode=0 Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.078435 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerDied","Data":"f81bcba504de02f46efd9f213fc1b9bf320f6ed872185ada8ac96c9e5d2ca74e"} Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.082201 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002" exitCode=0 Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.082282 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002"} Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.082525 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470"} Oct 13 12:55:56 crc kubenswrapper[4678]: I1013 12:55:56.082548 4678 scope.go:117] "RemoveContainer" containerID="13a2c321dc2cba92736c5c80e2a7461bc35ca9329b2343a88291061a4bfce44d" Oct 13 12:55:57 crc kubenswrapper[4678]: I1013 12:55:57.097767 4678 generic.go:334] "Generic (PLEG): container finished" podID="ead738f8-1d8c-491d-9c14-8cf4b535badf" containerID="0dc85a15269bf65975d949936dc76863d1ed3bdcb6bef9d5eeb655701f873baf" exitCode=0 Oct 13 12:55:57 crc kubenswrapper[4678]: I1013 12:55:57.097840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerDied","Data":"0dc85a15269bf65975d949936dc76863d1ed3bdcb6bef9d5eeb655701f873baf"} Oct 13 12:55:57 crc kubenswrapper[4678]: I1013 12:55:57.398820 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-x9tg5" Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.110564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"7538227ec46b2b8c262665097bcfb10999ed2908b5e8001b3ff518a4ba852ab0"} Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.110847 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"abf42758fa64c4f5d8d34e4850fd318724086832fc288a0ac02477416f77b7e6"} Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.110857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"a4ad3f8f0442abf6d1d17bdea832a00c15353d88b27cb5455189d90119187f1b"} Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.110866 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"12d6294f2396bf41fff33230a3d6ff719d25e7ff9ca8d46754d9dbca7f11c0f3"} Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.110874 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"01033322e8c3e33e37b6ba5ee212d4e9b28e8a987a5bfa40a3a4c256696ad7ab"} Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.578837 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mbsh9" Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.621859 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.622084 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" podUID="83521d15-1151-43e4-bd62-8aca811169dc" containerName="controller-manager" containerID="cri-o://a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a" gracePeriod=30 Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.699259 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:55:58 crc kubenswrapper[4678]: I1013 12:55:58.699467 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" podUID="7e814849-60c6-4107-8ed6-5301f9cda527" containerName="route-controller-manager" containerID="cri-o://739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800" gracePeriod=30 Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.045040 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.112006 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.115899 4678 generic.go:334] "Generic (PLEG): container finished" podID="83521d15-1151-43e4-bd62-8aca811169dc" containerID="a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a" exitCode=0 Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.115945 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.115968 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" event={"ID":"83521d15-1151-43e4-bd62-8aca811169dc","Type":"ContainerDied","Data":"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a"} Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.115996 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-s8h75" event={"ID":"83521d15-1151-43e4-bd62-8aca811169dc","Type":"ContainerDied","Data":"1280e4b0eb3a0a05adf01f5be44757766d8e5f877727091d4950e30f2708b060"} Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.116022 4678 scope.go:117] "RemoveContainer" containerID="a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.118653 4678 generic.go:334] "Generic (PLEG): container finished" podID="7e814849-60c6-4107-8ed6-5301f9cda527" containerID="739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800" exitCode=0 Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.118705 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.118744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" event={"ID":"7e814849-60c6-4107-8ed6-5301f9cda527","Type":"ContainerDied","Data":"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800"} Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.118772 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h" event={"ID":"7e814849-60c6-4107-8ed6-5301f9cda527","Type":"ContainerDied","Data":"ed83ec74a888daf7fe582b713b88897972b35faf42a11c6e3bc91ed9d2270b4d"} Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.125223 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hdzf4" event={"ID":"ead738f8-1d8c-491d-9c14-8cf4b535badf","Type":"ContainerStarted","Data":"eb3d951eb20967fc4faa2732b9f8a2e3856a9184350326b90751c9d7ffbd2636"} Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.125448 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.137620 4678 scope.go:117] "RemoveContainer" containerID="a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a" Oct 13 12:55:59 crc kubenswrapper[4678]: E1013 12:55:59.138092 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a\": container with ID starting with a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a not found: ID does not exist" containerID="a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.138144 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a"} err="failed to get container status \"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a\": rpc error: code = NotFound desc = could not find container \"a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a\": container with ID starting with a2c9e8a5985c33e08dd83555242664aad103e0edfb95e678ee0ce8fb555abc6a not found: ID does not exist" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.138176 4678 scope.go:117] "RemoveContainer" containerID="739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.167426 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config\") pod \"83521d15-1151-43e4-bd62-8aca811169dc\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.167466 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v82hd\" (UniqueName: \"kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd\") pod \"83521d15-1151-43e4-bd62-8aca811169dc\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.167492 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca\") pod \"7e814849-60c6-4107-8ed6-5301f9cda527\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.167509 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca\") pod \"83521d15-1151-43e4-bd62-8aca811169dc\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168375 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca" (OuterVolumeSpecName: "client-ca") pod "83521d15-1151-43e4-bd62-8aca811169dc" (UID: "83521d15-1151-43e4-bd62-8aca811169dc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168448 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca" (OuterVolumeSpecName: "client-ca") pod "7e814849-60c6-4107-8ed6-5301f9cda527" (UID: "7e814849-60c6-4107-8ed6-5301f9cda527"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168466 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config" (OuterVolumeSpecName: "config") pod "83521d15-1151-43e4-bd62-8aca811169dc" (UID: "83521d15-1151-43e4-bd62-8aca811169dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168598 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crbcn\" (UniqueName: \"kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn\") pod \"7e814849-60c6-4107-8ed6-5301f9cda527\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168633 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert\") pod \"83521d15-1151-43e4-bd62-8aca811169dc\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168651 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config\") pod \"7e814849-60c6-4107-8ed6-5301f9cda527\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168670 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert\") pod \"7e814849-60c6-4107-8ed6-5301f9cda527\" (UID: \"7e814849-60c6-4107-8ed6-5301f9cda527\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168688 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles\") pod \"83521d15-1151-43e4-bd62-8aca811169dc\" (UID: \"83521d15-1151-43e4-bd62-8aca811169dc\") " Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168897 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168914 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.168923 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-client-ca\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.171778 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config" (OuterVolumeSpecName: "config") pod "7e814849-60c6-4107-8ed6-5301f9cda527" (UID: "7e814849-60c6-4107-8ed6-5301f9cda527"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.172621 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "83521d15-1151-43e4-bd62-8aca811169dc" (UID: "83521d15-1151-43e4-bd62-8aca811169dc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.173521 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7e814849-60c6-4107-8ed6-5301f9cda527" (UID: "7e814849-60c6-4107-8ed6-5301f9cda527"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.173819 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd" (OuterVolumeSpecName: "kube-api-access-v82hd") pod "83521d15-1151-43e4-bd62-8aca811169dc" (UID: "83521d15-1151-43e4-bd62-8aca811169dc"). InnerVolumeSpecName "kube-api-access-v82hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.177014 4678 scope.go:117] "RemoveContainer" containerID="739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800" Oct 13 12:55:59 crc kubenswrapper[4678]: E1013 12:55:59.177520 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800\": container with ID starting with 739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800 not found: ID does not exist" containerID="739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.177553 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800"} err="failed to get container status \"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800\": rpc error: code = NotFound desc = could not find container \"739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800\": container with ID starting with 739fb5595ff202e566909c055fc1c8f21085ab2f7e14c8e05ae95ad7c2ae8800 not found: ID does not exist" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.177653 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn" (OuterVolumeSpecName: "kube-api-access-crbcn") pod "7e814849-60c6-4107-8ed6-5301f9cda527" (UID: "7e814849-60c6-4107-8ed6-5301f9cda527"). InnerVolumeSpecName "kube-api-access-crbcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.177897 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "83521d15-1151-43e4-bd62-8aca811169dc" (UID: "83521d15-1151-43e4-bd62-8aca811169dc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269665 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v82hd\" (UniqueName: \"kubernetes.io/projected/83521d15-1151-43e4-bd62-8aca811169dc-kube-api-access-v82hd\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269695 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crbcn\" (UniqueName: \"kubernetes.io/projected/7e814849-60c6-4107-8ed6-5301f9cda527-kube-api-access-crbcn\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269704 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83521d15-1151-43e4-bd62-8aca811169dc-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269714 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e814849-60c6-4107-8ed6-5301f9cda527-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269723 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e814849-60c6-4107-8ed6-5301f9cda527-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.269734 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/83521d15-1151-43e4-bd62-8aca811169dc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.444255 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hdzf4" podStartSLOduration=6.271800814 podStartE2EDuration="13.444225007s" podCreationTimestamp="2025-10-13 12:55:46 +0000 UTC" firstStartedPulling="2025-10-13 12:55:46.861238742 +0000 UTC m=+734.945776636" lastFinishedPulling="2025-10-13 12:55:54.033662935 +0000 UTC m=+742.118200829" observedRunningTime="2025-10-13 12:55:59.157205705 +0000 UTC m=+747.241743589" watchObservedRunningTime="2025-10-13 12:55:59.444225007 +0000 UTC m=+747.528762931" Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.448464 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.464418 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-s8h75"] Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.474676 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:55:59 crc kubenswrapper[4678]: I1013 12:55:59.481138 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-45q2h"] Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.544569 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42"] Oct 13 12:56:00 crc kubenswrapper[4678]: E1013 12:56:00.545616 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83521d15-1151-43e4-bd62-8aca811169dc" containerName="controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.545640 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="83521d15-1151-43e4-bd62-8aca811169dc" containerName="controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: E1013 12:56:00.545660 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e814849-60c6-4107-8ed6-5301f9cda527" containerName="route-controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.545671 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e814849-60c6-4107-8ed6-5301f9cda527" containerName="route-controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.545871 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e814849-60c6-4107-8ed6-5301f9cda527" containerName="route-controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.545897 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="83521d15-1151-43e4-bd62-8aca811169dc" containerName="controller-manager" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.546566 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.547119 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6df86b6b97-bqrgp"] Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.547785 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.549960 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551434 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551496 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551503 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551566 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551655 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.551718 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.552310 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.552342 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.552550 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.553047 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.553082 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.569351 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42"] Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.571538 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df86b6b97-bqrgp"] Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.572637 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586496 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/854828a8-928d-479e-a79d-268d56bb1b51-serving-cert\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586547 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-proxy-ca-bundles\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586576 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db244068-1569-4bc4-a64d-b358c475f7b6-serving-cert\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586630 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-config\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586681 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-client-ca\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586713 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-config\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndlgs\" (UniqueName: \"kubernetes.io/projected/db244068-1569-4bc4-a64d-b358c475f7b6-kube-api-access-ndlgs\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btlq5\" (UniqueName: \"kubernetes.io/projected/854828a8-928d-479e-a79d-268d56bb1b51-kube-api-access-btlq5\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.586813 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-client-ca\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.603427 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e814849-60c6-4107-8ed6-5301f9cda527" path="/var/lib/kubelet/pods/7e814849-60c6-4107-8ed6-5301f9cda527/volumes" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.604259 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83521d15-1151-43e4-bd62-8aca811169dc" path="/var/lib/kubelet/pods/83521d15-1151-43e4-bd62-8aca811169dc/volumes" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.687760 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-proxy-ca-bundles\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.687814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db244068-1569-4bc4-a64d-b358c475f7b6-serving-cert\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.687844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-config\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.687892 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-client-ca\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.689133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-config\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.689200 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-config\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.689229 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndlgs\" (UniqueName: \"kubernetes.io/projected/db244068-1569-4bc4-a64d-b358c475f7b6-kube-api-access-ndlgs\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.689269 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-proxy-ca-bundles\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.689679 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/854828a8-928d-479e-a79d-268d56bb1b51-client-ca\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.690271 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btlq5\" (UniqueName: \"kubernetes.io/projected/854828a8-928d-479e-a79d-268d56bb1b51-kube-api-access-btlq5\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.690316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-client-ca\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.690352 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/854828a8-928d-479e-a79d-268d56bb1b51-serving-cert\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.690360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-config\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.691225 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/db244068-1569-4bc4-a64d-b358c475f7b6-client-ca\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.706320 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/854828a8-928d-479e-a79d-268d56bb1b51-serving-cert\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.706441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/db244068-1569-4bc4-a64d-b358c475f7b6-serving-cert\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.708783 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btlq5\" (UniqueName: \"kubernetes.io/projected/854828a8-928d-479e-a79d-268d56bb1b51-kube-api-access-btlq5\") pod \"route-controller-manager-cd5887bd5-snp42\" (UID: \"854828a8-928d-479e-a79d-268d56bb1b51\") " pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.710027 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndlgs\" (UniqueName: \"kubernetes.io/projected/db244068-1569-4bc4-a64d-b358c475f7b6-kube-api-access-ndlgs\") pod \"controller-manager-6df86b6b97-bqrgp\" (UID: \"db244068-1569-4bc4-a64d-b358c475f7b6\") " pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.871639 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:00 crc kubenswrapper[4678]: I1013 12:56:00.881132 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.105164 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6df86b6b97-bqrgp"] Oct 13 12:56:01 crc kubenswrapper[4678]: W1013 12:56:01.108043 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb244068_1569_4bc4_a64d_b358c475f7b6.slice/crio-2cb1960a1a4ac918ca177c0c7d54fa968c6a7994d7a6f80d38745ff5cc8ad461 WatchSource:0}: Error finding container 2cb1960a1a4ac918ca177c0c7d54fa968c6a7994d7a6f80d38745ff5cc8ad461: Status 404 returned error can't find the container with id 2cb1960a1a4ac918ca177c0c7d54fa968c6a7994d7a6f80d38745ff5cc8ad461 Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.146202 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" event={"ID":"db244068-1569-4bc4-a64d-b358c475f7b6","Type":"ContainerStarted","Data":"2cb1960a1a4ac918ca177c0c7d54fa968c6a7994d7a6f80d38745ff5cc8ad461"} Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.373212 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42"] Oct 13 12:56:01 crc kubenswrapper[4678]: W1013 12:56:01.378949 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod854828a8_928d_479e_a79d_268d56bb1b51.slice/crio-43258345d3d680e4d56649583c9213a31001f753f064936ed7ff1fd0ef12a0e3 WatchSource:0}: Error finding container 43258345d3d680e4d56649583c9213a31001f753f064936ed7ff1fd0ef12a0e3: Status 404 returned error can't find the container with id 43258345d3d680e4d56649583c9213a31001f753f064936ed7ff1fd0ef12a0e3 Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.705882 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.762806 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.763518 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.765805 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.766138 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.766150 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.776866 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.803778 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8tm\" (UniqueName: \"kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm\") pod \"openstack-operator-index-v9fmj\" (UID: \"9a43d470-d814-4a27-ae90-b96c679ff903\") " pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.904620 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8tm\" (UniqueName: \"kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm\") pod \"openstack-operator-index-v9fmj\" (UID: \"9a43d470-d814-4a27-ae90-b96c679ff903\") " pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:01 crc kubenswrapper[4678]: I1013 12:56:01.922867 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8tm\" (UniqueName: \"kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm\") pod \"openstack-operator-index-v9fmj\" (UID: \"9a43d470-d814-4a27-ae90-b96c679ff903\") " pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.077861 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.157157 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" event={"ID":"854828a8-928d-479e-a79d-268d56bb1b51","Type":"ContainerStarted","Data":"ad91e06b17ef781ccdd088c7611e1b5d3392e0084470c9e510719379e0743e29"} Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.157216 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" event={"ID":"854828a8-928d-479e-a79d-268d56bb1b51","Type":"ContainerStarted","Data":"43258345d3d680e4d56649583c9213a31001f753f064936ed7ff1fd0ef12a0e3"} Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.157506 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.161140 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" event={"ID":"db244068-1569-4bc4-a64d-b358c475f7b6","Type":"ContainerStarted","Data":"6821a4eeef06dfc31ae4dfb67003d470148940bfd49202079f1bc82a5eaa4e10"} Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.161181 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.162202 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.165594 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.173877 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cd5887bd5-snp42" podStartSLOduration=4.173859954 podStartE2EDuration="4.173859954s" podCreationTimestamp="2025-10-13 12:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:56:02.170848695 +0000 UTC m=+750.255386579" watchObservedRunningTime="2025-10-13 12:56:02.173859954 +0000 UTC m=+750.258397838" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.263042 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6df86b6b97-bqrgp" podStartSLOduration=4.26302579 podStartE2EDuration="4.26302579s" podCreationTimestamp="2025-10-13 12:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:56:02.258433809 +0000 UTC m=+750.342971693" watchObservedRunningTime="2025-10-13 12:56:02.26302579 +0000 UTC m=+750.347563674" Oct 13 12:56:02 crc kubenswrapper[4678]: I1013 12:56:02.563494 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:02 crc kubenswrapper[4678]: W1013 12:56:02.566641 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a43d470_d814_4a27_ae90_b96c679ff903.slice/crio-2399ced323786a629bc31adeae845751ca0e20646a5bfd0f870563f04874c1ce WatchSource:0}: Error finding container 2399ced323786a629bc31adeae845751ca0e20646a5bfd0f870563f04874c1ce: Status 404 returned error can't find the container with id 2399ced323786a629bc31adeae845751ca0e20646a5bfd0f870563f04874c1ce Oct 13 12:56:03 crc kubenswrapper[4678]: I1013 12:56:03.168223 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v9fmj" event={"ID":"9a43d470-d814-4a27-ae90-b96c679ff903","Type":"ContainerStarted","Data":"2399ced323786a629bc31adeae845751ca0e20646a5bfd0f870563f04874c1ce"} Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.125399 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.735758 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4fn9q"] Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.736588 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.739926 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xwwcg" Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.743842 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4fn9q"] Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.761542 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4rd\" (UniqueName: \"kubernetes.io/projected/dadbd9ec-7f53-4555-951c-2d090c017417-kube-api-access-mk4rd\") pod \"openstack-operator-index-4fn9q\" (UID: \"dadbd9ec-7f53-4555-951c-2d090c017417\") " pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.862793 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4rd\" (UniqueName: \"kubernetes.io/projected/dadbd9ec-7f53-4555-951c-2d090c017417-kube-api-access-mk4rd\") pod \"openstack-operator-index-4fn9q\" (UID: \"dadbd9ec-7f53-4555-951c-2d090c017417\") " pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:05 crc kubenswrapper[4678]: I1013 12:56:05.885849 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4rd\" (UniqueName: \"kubernetes.io/projected/dadbd9ec-7f53-4555-951c-2d090c017417-kube-api-access-mk4rd\") pod \"openstack-operator-index-4fn9q\" (UID: \"dadbd9ec-7f53-4555-951c-2d090c017417\") " pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.054576 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.193161 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v9fmj" event={"ID":"9a43d470-d814-4a27-ae90-b96c679ff903","Type":"ContainerStarted","Data":"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d"} Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.193316 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-v9fmj" podUID="9a43d470-d814-4a27-ae90-b96c679ff903" containerName="registry-server" containerID="cri-o://cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d" gracePeriod=2 Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.217136 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-v9fmj" podStartSLOduration=2.064350097 podStartE2EDuration="5.215249077s" podCreationTimestamp="2025-10-13 12:56:01 +0000 UTC" firstStartedPulling="2025-10-13 12:56:02.569342931 +0000 UTC m=+750.653880815" lastFinishedPulling="2025-10-13 12:56:05.720241911 +0000 UTC m=+753.804779795" observedRunningTime="2025-10-13 12:56:06.212956956 +0000 UTC m=+754.297494860" watchObservedRunningTime="2025-10-13 12:56:06.215249077 +0000 UTC m=+754.299786971" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.515804 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4fn9q"] Oct 13 12:56:06 crc kubenswrapper[4678]: W1013 12:56:06.522464 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddadbd9ec_7f53_4555_951c_2d090c017417.slice/crio-b86fd0d162611b6ae7e3d002eb7fee5eb699545f2c415f495f33ff79f88a8f9d WatchSource:0}: Error finding container b86fd0d162611b6ae7e3d002eb7fee5eb699545f2c415f495f33ff79f88a8f9d: Status 404 returned error can't find the container with id b86fd0d162611b6ae7e3d002eb7fee5eb699545f2c415f495f33ff79f88a8f9d Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.597199 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.671750 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj8tm\" (UniqueName: \"kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm\") pod \"9a43d470-d814-4a27-ae90-b96c679ff903\" (UID: \"9a43d470-d814-4a27-ae90-b96c679ff903\") " Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.677553 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm" (OuterVolumeSpecName: "kube-api-access-wj8tm") pod "9a43d470-d814-4a27-ae90-b96c679ff903" (UID: "9a43d470-d814-4a27-ae90-b96c679ff903"). InnerVolumeSpecName "kube-api-access-wj8tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.698931 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-xp775" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.773690 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj8tm\" (UniqueName: \"kubernetes.io/projected/9a43d470-d814-4a27-ae90-b96c679ff903-kube-api-access-wj8tm\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:06 crc kubenswrapper[4678]: I1013 12:56:06.969985 4678 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.200473 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4fn9q" event={"ID":"dadbd9ec-7f53-4555-951c-2d090c017417","Type":"ContainerStarted","Data":"106511cce8d99a25ad1fd5d6f3d5c1627ca166a946cef651a5919473660eb3a6"} Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.200516 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4fn9q" event={"ID":"dadbd9ec-7f53-4555-951c-2d090c017417","Type":"ContainerStarted","Data":"b86fd0d162611b6ae7e3d002eb7fee5eb699545f2c415f495f33ff79f88a8f9d"} Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.201883 4678 generic.go:334] "Generic (PLEG): container finished" podID="9a43d470-d814-4a27-ae90-b96c679ff903" containerID="cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d" exitCode=0 Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.201912 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v9fmj" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.201930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v9fmj" event={"ID":"9a43d470-d814-4a27-ae90-b96c679ff903","Type":"ContainerDied","Data":"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d"} Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.201960 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v9fmj" event={"ID":"9a43d470-d814-4a27-ae90-b96c679ff903","Type":"ContainerDied","Data":"2399ced323786a629bc31adeae845751ca0e20646a5bfd0f870563f04874c1ce"} Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.201982 4678 scope.go:117] "RemoveContainer" containerID="cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.225579 4678 scope.go:117] "RemoveContainer" containerID="cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d" Oct 13 12:56:07 crc kubenswrapper[4678]: E1013 12:56:07.229917 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d\": container with ID starting with cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d not found: ID does not exist" containerID="cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.229958 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d"} err="failed to get container status \"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d\": rpc error: code = NotFound desc = could not find container \"cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d\": container with ID starting with cd6d39ecfb532614551231eb43fbb1e1b9c498935c405e28d2e6fafde8ef092d not found: ID does not exist" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.231841 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4fn9q" podStartSLOduration=2.173237855 podStartE2EDuration="2.231804216s" podCreationTimestamp="2025-10-13 12:56:05 +0000 UTC" firstStartedPulling="2025-10-13 12:56:06.52850941 +0000 UTC m=+754.613047304" lastFinishedPulling="2025-10-13 12:56:06.587075781 +0000 UTC m=+754.671613665" observedRunningTime="2025-10-13 12:56:07.225805928 +0000 UTC m=+755.310343812" watchObservedRunningTime="2025-10-13 12:56:07.231804216 +0000 UTC m=+755.316342140" Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.241391 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:07 crc kubenswrapper[4678]: I1013 12:56:07.243805 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-v9fmj"] Oct 13 12:56:08 crc kubenswrapper[4678]: I1013 12:56:08.604034 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a43d470-d814-4a27-ae90-b96c679ff903" path="/var/lib/kubelet/pods/9a43d470-d814-4a27-ae90-b96c679ff903/volumes" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.347098 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:10 crc kubenswrapper[4678]: E1013 12:56:10.347563 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a43d470-d814-4a27-ae90-b96c679ff903" containerName="registry-server" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.347574 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a43d470-d814-4a27-ae90-b96c679ff903" containerName="registry-server" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.347694 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a43d470-d814-4a27-ae90-b96c679ff903" containerName="registry-server" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.348519 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.380893 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.419783 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.419825 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.419903 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdxwd\" (UniqueName: \"kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.521373 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.521417 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.521486 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdxwd\" (UniqueName: \"kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.521931 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.522041 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.543358 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdxwd\" (UniqueName: \"kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd\") pod \"redhat-operators-td54x\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:10 crc kubenswrapper[4678]: I1013 12:56:10.665565 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:11 crc kubenswrapper[4678]: I1013 12:56:11.070363 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:11 crc kubenswrapper[4678]: W1013 12:56:11.078722 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod132c11d0_9cb6_4bba_a8f5_9575c631ecd2.slice/crio-2ca2ca78f4e399009730364073d434415f214dfdabfafb020f2b825448dc02e9 WatchSource:0}: Error finding container 2ca2ca78f4e399009730364073d434415f214dfdabfafb020f2b825448dc02e9: Status 404 returned error can't find the container with id 2ca2ca78f4e399009730364073d434415f214dfdabfafb020f2b825448dc02e9 Oct 13 12:56:11 crc kubenswrapper[4678]: I1013 12:56:11.225417 4678 generic.go:334] "Generic (PLEG): container finished" podID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerID="648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc" exitCode=0 Oct 13 12:56:11 crc kubenswrapper[4678]: I1013 12:56:11.225488 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerDied","Data":"648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc"} Oct 13 12:56:11 crc kubenswrapper[4678]: I1013 12:56:11.225734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerStarted","Data":"2ca2ca78f4e399009730364073d434415f214dfdabfafb020f2b825448dc02e9"} Oct 13 12:56:13 crc kubenswrapper[4678]: I1013 12:56:13.242141 4678 generic.go:334] "Generic (PLEG): container finished" podID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerID="0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce" exitCode=0 Oct 13 12:56:13 crc kubenswrapper[4678]: I1013 12:56:13.242239 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerDied","Data":"0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce"} Oct 13 12:56:14 crc kubenswrapper[4678]: I1013 12:56:14.255900 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerStarted","Data":"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210"} Oct 13 12:56:14 crc kubenswrapper[4678]: I1013 12:56:14.287023 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-td54x" podStartSLOduration=1.516604915 podStartE2EDuration="4.286995185s" podCreationTimestamp="2025-10-13 12:56:10 +0000 UTC" firstStartedPulling="2025-10-13 12:56:11.226522792 +0000 UTC m=+759.311060676" lastFinishedPulling="2025-10-13 12:56:13.996913012 +0000 UTC m=+762.081450946" observedRunningTime="2025-10-13 12:56:14.283850492 +0000 UTC m=+762.368388406" watchObservedRunningTime="2025-10-13 12:56:14.286995185 +0000 UTC m=+762.371533079" Oct 13 12:56:16 crc kubenswrapper[4678]: I1013 12:56:16.054948 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:16 crc kubenswrapper[4678]: I1013 12:56:16.055426 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:16 crc kubenswrapper[4678]: I1013 12:56:16.104456 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:16 crc kubenswrapper[4678]: I1013 12:56:16.333484 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4fn9q" Oct 13 12:56:16 crc kubenswrapper[4678]: I1013 12:56:16.711264 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hdzf4" Oct 13 12:56:20 crc kubenswrapper[4678]: I1013 12:56:20.666372 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:20 crc kubenswrapper[4678]: I1013 12:56:20.667948 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:20 crc kubenswrapper[4678]: I1013 12:56:20.731708 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.392007 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.566291 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m"] Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.567409 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.570230 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-h9fmx" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.578950 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m"] Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.673138 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.673601 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.673734 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h6f5\" (UniqueName: \"kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.775626 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.776385 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.776534 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h6f5\" (UniqueName: \"kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.776391 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.776808 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.803223 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h6f5\" (UniqueName: \"kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5\") pod \"846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:21 crc kubenswrapper[4678]: I1013 12:56:21.898853 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.131941 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.136543 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.152876 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.182254 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.182400 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.182428 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcc56\" (UniqueName: \"kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.283728 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.283772 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcc56\" (UniqueName: \"kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.283822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.284284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.284403 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.303085 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcc56\" (UniqueName: \"kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56\") pod \"redhat-marketplace-jvb28\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.344742 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m"] Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.456119 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:22 crc kubenswrapper[4678]: I1013 12:56:22.865187 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:22 crc kubenswrapper[4678]: W1013 12:56:22.867265 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bdede69_03d1_4981_bf7b_377846abbfd7.slice/crio-bde387c7d8e48ddba3a83cc7f1172401aefad8a5c7312ddfd66798cff254145b WatchSource:0}: Error finding container bde387c7d8e48ddba3a83cc7f1172401aefad8a5c7312ddfd66798cff254145b: Status 404 returned error can't find the container with id bde387c7d8e48ddba3a83cc7f1172401aefad8a5c7312ddfd66798cff254145b Oct 13 12:56:23 crc kubenswrapper[4678]: I1013 12:56:23.333758 4678 generic.go:334] "Generic (PLEG): container finished" podID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerID="7c5c1a5060ecd7d8957a92e82b7afaed64cfb1a6c7c97e229675b6e846c1a69c" exitCode=0 Oct 13 12:56:23 crc kubenswrapper[4678]: I1013 12:56:23.333844 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" event={"ID":"0908157e-2110-4b3d-baa9-92a9d8c7a3c4","Type":"ContainerDied","Data":"7c5c1a5060ecd7d8957a92e82b7afaed64cfb1a6c7c97e229675b6e846c1a69c"} Oct 13 12:56:23 crc kubenswrapper[4678]: I1013 12:56:23.334207 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" event={"ID":"0908157e-2110-4b3d-baa9-92a9d8c7a3c4","Type":"ContainerStarted","Data":"ae46f68f0a1de991e7c6071f0ce1b41fd8b1b1e95c6e3de2c25b07e2a73f9afe"} Oct 13 12:56:23 crc kubenswrapper[4678]: I1013 12:56:23.340302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerStarted","Data":"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b"} Oct 13 12:56:23 crc kubenswrapper[4678]: I1013 12:56:23.340393 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerStarted","Data":"bde387c7d8e48ddba3a83cc7f1172401aefad8a5c7312ddfd66798cff254145b"} Oct 13 12:56:24 crc kubenswrapper[4678]: I1013 12:56:24.352800 4678 generic.go:334] "Generic (PLEG): container finished" podID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerID="1e4b75a46771c44ea45d216a034c69d759f38c71445d561099acfee62ba232f6" exitCode=0 Oct 13 12:56:24 crc kubenswrapper[4678]: I1013 12:56:24.352915 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" event={"ID":"0908157e-2110-4b3d-baa9-92a9d8c7a3c4","Type":"ContainerDied","Data":"1e4b75a46771c44ea45d216a034c69d759f38c71445d561099acfee62ba232f6"} Oct 13 12:56:24 crc kubenswrapper[4678]: I1013 12:56:24.355317 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerID="6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b" exitCode=0 Oct 13 12:56:24 crc kubenswrapper[4678]: I1013 12:56:24.355383 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerDied","Data":"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b"} Oct 13 12:56:25 crc kubenswrapper[4678]: I1013 12:56:25.372542 4678 generic.go:334] "Generic (PLEG): container finished" podID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerID="aa526ad7199e45cab54e96c142e65db8bd7c2165227b636a2ba4a7167bee2a34" exitCode=0 Oct 13 12:56:25 crc kubenswrapper[4678]: I1013 12:56:25.372620 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" event={"ID":"0908157e-2110-4b3d-baa9-92a9d8c7a3c4","Type":"ContainerDied","Data":"aa526ad7199e45cab54e96c142e65db8bd7c2165227b636a2ba4a7167bee2a34"} Oct 13 12:56:25 crc kubenswrapper[4678]: I1013 12:56:25.533670 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:25 crc kubenswrapper[4678]: I1013 12:56:25.533904 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-td54x" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="registry-server" containerID="cri-o://969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210" gracePeriod=2 Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.068772 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.147181 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities\") pod \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.147265 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content\") pod \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.147326 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdxwd\" (UniqueName: \"kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd\") pod \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\" (UID: \"132c11d0-9cb6-4bba-a8f5-9575c631ecd2\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.157080 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities" (OuterVolumeSpecName: "utilities") pod "132c11d0-9cb6-4bba-a8f5-9575c631ecd2" (UID: "132c11d0-9cb6-4bba-a8f5-9575c631ecd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.169224 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd" (OuterVolumeSpecName: "kube-api-access-kdxwd") pod "132c11d0-9cb6-4bba-a8f5-9575c631ecd2" (UID: "132c11d0-9cb6-4bba-a8f5-9575c631ecd2"). InnerVolumeSpecName "kube-api-access-kdxwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.227536 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "132c11d0-9cb6-4bba-a8f5-9575c631ecd2" (UID: "132c11d0-9cb6-4bba-a8f5-9575c631ecd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.248768 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdxwd\" (UniqueName: \"kubernetes.io/projected/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-kube-api-access-kdxwd\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.248814 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.248831 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/132c11d0-9cb6-4bba-a8f5-9575c631ecd2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.383123 4678 generic.go:334] "Generic (PLEG): container finished" podID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerID="969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210" exitCode=0 Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.383205 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerDied","Data":"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210"} Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.383229 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-td54x" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.383257 4678 scope.go:117] "RemoveContainer" containerID="969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.383240 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-td54x" event={"ID":"132c11d0-9cb6-4bba-a8f5-9575c631ecd2","Type":"ContainerDied","Data":"2ca2ca78f4e399009730364073d434415f214dfdabfafb020f2b825448dc02e9"} Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.386316 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerID="c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc" exitCode=0 Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.387366 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerDied","Data":"c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc"} Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.402697 4678 scope.go:117] "RemoveContainer" containerID="0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.427310 4678 scope.go:117] "RemoveContainer" containerID="648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.434558 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.443884 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-td54x"] Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.462801 4678 scope.go:117] "RemoveContainer" containerID="969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210" Oct 13 12:56:26 crc kubenswrapper[4678]: E1013 12:56:26.463374 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210\": container with ID starting with 969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210 not found: ID does not exist" containerID="969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.463408 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210"} err="failed to get container status \"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210\": rpc error: code = NotFound desc = could not find container \"969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210\": container with ID starting with 969fcbb3257993b45f1841fd246be1fb1cf192a84866a130bf7090ec76a20210 not found: ID does not exist" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.463426 4678 scope.go:117] "RemoveContainer" containerID="0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce" Oct 13 12:56:26 crc kubenswrapper[4678]: E1013 12:56:26.467964 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce\": container with ID starting with 0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce not found: ID does not exist" containerID="0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.467989 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce"} err="failed to get container status \"0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce\": rpc error: code = NotFound desc = could not find container \"0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce\": container with ID starting with 0a63e0ef42f9c6b1a324e4c3b445a495ef01151dde488ed8c108d5e10af611ce not found: ID does not exist" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.468007 4678 scope.go:117] "RemoveContainer" containerID="648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc" Oct 13 12:56:26 crc kubenswrapper[4678]: E1013 12:56:26.469684 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc\": container with ID starting with 648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc not found: ID does not exist" containerID="648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.469707 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc"} err="failed to get container status \"648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc\": rpc error: code = NotFound desc = could not find container \"648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc\": container with ID starting with 648740fe4fabcfe7e1309c5a9dbb4a90c8272def5e309a65dd06246ad94a1bdc not found: ID does not exist" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.603077 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" path="/var/lib/kubelet/pods/132c11d0-9cb6-4bba-a8f5-9575c631ecd2/volumes" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.781492 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.856588 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle\") pod \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.856660 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util\") pod \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.856768 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h6f5\" (UniqueName: \"kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5\") pod \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\" (UID: \"0908157e-2110-4b3d-baa9-92a9d8c7a3c4\") " Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.857843 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle" (OuterVolumeSpecName: "bundle") pod "0908157e-2110-4b3d-baa9-92a9d8c7a3c4" (UID: "0908157e-2110-4b3d-baa9-92a9d8c7a3c4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.863315 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5" (OuterVolumeSpecName: "kube-api-access-4h6f5") pod "0908157e-2110-4b3d-baa9-92a9d8c7a3c4" (UID: "0908157e-2110-4b3d-baa9-92a9d8c7a3c4"). InnerVolumeSpecName "kube-api-access-4h6f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.878618 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util" (OuterVolumeSpecName: "util") pod "0908157e-2110-4b3d-baa9-92a9d8c7a3c4" (UID: "0908157e-2110-4b3d-baa9-92a9d8c7a3c4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.958989 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.959072 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-util\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:26 crc kubenswrapper[4678]: I1013 12:56:26.959108 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h6f5\" (UniqueName: \"kubernetes.io/projected/0908157e-2110-4b3d-baa9-92a9d8c7a3c4-kube-api-access-4h6f5\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:27 crc kubenswrapper[4678]: I1013 12:56:27.401484 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" event={"ID":"0908157e-2110-4b3d-baa9-92a9d8c7a3c4","Type":"ContainerDied","Data":"ae46f68f0a1de991e7c6071f0ce1b41fd8b1b1e95c6e3de2c25b07e2a73f9afe"} Oct 13 12:56:27 crc kubenswrapper[4678]: I1013 12:56:27.401857 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae46f68f0a1de991e7c6071f0ce1b41fd8b1b1e95c6e3de2c25b07e2a73f9afe" Oct 13 12:56:27 crc kubenswrapper[4678]: I1013 12:56:27.401997 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m" Oct 13 12:56:27 crc kubenswrapper[4678]: I1013 12:56:27.408016 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerStarted","Data":"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604"} Oct 13 12:56:27 crc kubenswrapper[4678]: I1013 12:56:27.434551 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jvb28" podStartSLOduration=2.979603575 podStartE2EDuration="5.434524364s" podCreationTimestamp="2025-10-13 12:56:22 +0000 UTC" firstStartedPulling="2025-10-13 12:56:24.358681977 +0000 UTC m=+772.443219901" lastFinishedPulling="2025-10-13 12:56:26.813602816 +0000 UTC m=+774.898140690" observedRunningTime="2025-10-13 12:56:27.428874925 +0000 UTC m=+775.513412819" watchObservedRunningTime="2025-10-13 12:56:27.434524364 +0000 UTC m=+775.519062288" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929372 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv"] Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929857 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="extract" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929869 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="extract" Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929877 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="extract-utilities" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929883 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="extract-utilities" Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929894 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="extract-content" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929901 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="extract-content" Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929912 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="util" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929918 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="util" Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929928 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="pull" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929933 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="pull" Oct 13 12:56:29 crc kubenswrapper[4678]: E1013 12:56:29.929945 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="registry-server" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.929951 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="registry-server" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.930040 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="132c11d0-9cb6-4bba-a8f5-9575c631ecd2" containerName="registry-server" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.930066 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908157e-2110-4b3d-baa9-92a9d8c7a3c4" containerName="extract" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.930680 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.936009 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-gc72l" Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.957357 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv"] Oct 13 12:56:29 crc kubenswrapper[4678]: I1013 12:56:29.999877 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6jqg\" (UniqueName: \"kubernetes.io/projected/e6f72e4a-d252-4053-9665-db9af69e697e-kube-api-access-x6jqg\") pod \"openstack-operator-controller-operator-766c4fbc4c-ddcpv\" (UID: \"e6f72e4a-d252-4053-9665-db9af69e697e\") " pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:30 crc kubenswrapper[4678]: I1013 12:56:30.101039 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6jqg\" (UniqueName: \"kubernetes.io/projected/e6f72e4a-d252-4053-9665-db9af69e697e-kube-api-access-x6jqg\") pod \"openstack-operator-controller-operator-766c4fbc4c-ddcpv\" (UID: \"e6f72e4a-d252-4053-9665-db9af69e697e\") " pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:30 crc kubenswrapper[4678]: I1013 12:56:30.121754 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6jqg\" (UniqueName: \"kubernetes.io/projected/e6f72e4a-d252-4053-9665-db9af69e697e-kube-api-access-x6jqg\") pod \"openstack-operator-controller-operator-766c4fbc4c-ddcpv\" (UID: \"e6f72e4a-d252-4053-9665-db9af69e697e\") " pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:30 crc kubenswrapper[4678]: I1013 12:56:30.246927 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:30 crc kubenswrapper[4678]: I1013 12:56:30.678703 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv"] Oct 13 12:56:30 crc kubenswrapper[4678]: W1013 12:56:30.682215 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6f72e4a_d252_4053_9665_db9af69e697e.slice/crio-fe510eb096f36b794ac63407d699a0d070f0a53c86ebee31bd001ee2fcffbd3d WatchSource:0}: Error finding container fe510eb096f36b794ac63407d699a0d070f0a53c86ebee31bd001ee2fcffbd3d: Status 404 returned error can't find the container with id fe510eb096f36b794ac63407d699a0d070f0a53c86ebee31bd001ee2fcffbd3d Oct 13 12:56:31 crc kubenswrapper[4678]: I1013 12:56:31.438152 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" event={"ID":"e6f72e4a-d252-4053-9665-db9af69e697e","Type":"ContainerStarted","Data":"fe510eb096f36b794ac63407d699a0d070f0a53c86ebee31bd001ee2fcffbd3d"} Oct 13 12:56:32 crc kubenswrapper[4678]: I1013 12:56:32.457252 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:32 crc kubenswrapper[4678]: I1013 12:56:32.457596 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:32 crc kubenswrapper[4678]: I1013 12:56:32.519871 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:33 crc kubenswrapper[4678]: I1013 12:56:33.502510 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:34 crc kubenswrapper[4678]: I1013 12:56:34.455889 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" event={"ID":"e6f72e4a-d252-4053-9665-db9af69e697e","Type":"ContainerStarted","Data":"a078b6573b58acd6f75680539d86744af5c9a22ec62b98e62c15f2aa28af2a78"} Oct 13 12:56:35 crc kubenswrapper[4678]: I1013 12:56:35.128118 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:35 crc kubenswrapper[4678]: I1013 12:56:35.462881 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jvb28" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="registry-server" containerID="cri-o://3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604" gracePeriod=2 Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.380726 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.472890 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerID="3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604" exitCode=0 Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.472965 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerDied","Data":"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604"} Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.472977 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jvb28" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.473009 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jvb28" event={"ID":"4bdede69-03d1-4981-bf7b-377846abbfd7","Type":"ContainerDied","Data":"bde387c7d8e48ddba3a83cc7f1172401aefad8a5c7312ddfd66798cff254145b"} Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.473044 4678 scope.go:117] "RemoveContainer" containerID="3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.491412 4678 scope.go:117] "RemoveContainer" containerID="c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.505655 4678 scope.go:117] "RemoveContainer" containerID="6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.524502 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content\") pod \"4bdede69-03d1-4981-bf7b-377846abbfd7\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.525273 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities\") pod \"4bdede69-03d1-4981-bf7b-377846abbfd7\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.525342 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcc56\" (UniqueName: \"kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56\") pod \"4bdede69-03d1-4981-bf7b-377846abbfd7\" (UID: \"4bdede69-03d1-4981-bf7b-377846abbfd7\") " Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.525972 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities" (OuterVolumeSpecName: "utilities") pod "4bdede69-03d1-4981-bf7b-377846abbfd7" (UID: "4bdede69-03d1-4981-bf7b-377846abbfd7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.527929 4678 scope.go:117] "RemoveContainer" containerID="3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604" Oct 13 12:56:36 crc kubenswrapper[4678]: E1013 12:56:36.528358 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604\": container with ID starting with 3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604 not found: ID does not exist" containerID="3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.528397 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604"} err="failed to get container status \"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604\": rpc error: code = NotFound desc = could not find container \"3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604\": container with ID starting with 3d4a98c4cf0713de1467eb70efe3f343f5dd0259c222e39162757277a1769604 not found: ID does not exist" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.528423 4678 scope.go:117] "RemoveContainer" containerID="c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc" Oct 13 12:56:36 crc kubenswrapper[4678]: E1013 12:56:36.528955 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc\": container with ID starting with c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc not found: ID does not exist" containerID="c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.528980 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc"} err="failed to get container status \"c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc\": rpc error: code = NotFound desc = could not find container \"c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc\": container with ID starting with c36b4949994bebed767d6b134af6b560aba4f69587b617472c10088205531cfc not found: ID does not exist" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.528996 4678 scope.go:117] "RemoveContainer" containerID="6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b" Oct 13 12:56:36 crc kubenswrapper[4678]: E1013 12:56:36.529371 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b\": container with ID starting with 6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b not found: ID does not exist" containerID="6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.529390 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b"} err="failed to get container status \"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b\": rpc error: code = NotFound desc = could not find container \"6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b\": container with ID starting with 6e7686d307236f3aa2a2c2432228c48289330bcc5db5805862486c30a8af0f9b not found: ID does not exist" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.530731 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56" (OuterVolumeSpecName: "kube-api-access-hcc56") pod "4bdede69-03d1-4981-bf7b-377846abbfd7" (UID: "4bdede69-03d1-4981-bf7b-377846abbfd7"). InnerVolumeSpecName "kube-api-access-hcc56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.547465 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bdede69-03d1-4981-bf7b-377846abbfd7" (UID: "4bdede69-03d1-4981-bf7b-377846abbfd7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.626457 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.626482 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcc56\" (UniqueName: \"kubernetes.io/projected/4bdede69-03d1-4981-bf7b-377846abbfd7-kube-api-access-hcc56\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.626492 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdede69-03d1-4981-bf7b-377846abbfd7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.795886 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:36 crc kubenswrapper[4678]: I1013 12:56:36.799707 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jvb28"] Oct 13 12:56:37 crc kubenswrapper[4678]: I1013 12:56:37.483179 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" event={"ID":"e6f72e4a-d252-4053-9665-db9af69e697e","Type":"ContainerStarted","Data":"c6104d49d3dad36f40adb8b5754aff8c1c590b88e431357993f5580e8455cd95"} Oct 13 12:56:37 crc kubenswrapper[4678]: I1013 12:56:37.483284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:37 crc kubenswrapper[4678]: I1013 12:56:37.556264 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" podStartSLOduration=2.883000631 podStartE2EDuration="8.556226237s" podCreationTimestamp="2025-10-13 12:56:29 +0000 UTC" firstStartedPulling="2025-10-13 12:56:30.685384151 +0000 UTC m=+778.769922045" lastFinishedPulling="2025-10-13 12:56:36.358609757 +0000 UTC m=+784.443147651" observedRunningTime="2025-10-13 12:56:37.542718017 +0000 UTC m=+785.627255941" watchObservedRunningTime="2025-10-13 12:56:37.556226237 +0000 UTC m=+785.640764171" Oct 13 12:56:38 crc kubenswrapper[4678]: I1013 12:56:38.602725 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" path="/var/lib/kubelet/pods/4bdede69-03d1-4981-bf7b-377846abbfd7/volumes" Oct 13 12:56:40 crc kubenswrapper[4678]: I1013 12:56:40.250175 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-766c4fbc4c-ddcpv" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.642898 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5"] Oct 13 12:56:56 crc kubenswrapper[4678]: E1013 12:56:56.643948 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="registry-server" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.643965 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="registry-server" Oct 13 12:56:56 crc kubenswrapper[4678]: E1013 12:56:56.643977 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="extract-content" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.643984 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="extract-content" Oct 13 12:56:56 crc kubenswrapper[4678]: E1013 12:56:56.643996 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="extract-utilities" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.644002 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="extract-utilities" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.644149 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdede69-03d1-4981-bf7b-377846abbfd7" containerName="registry-server" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.644903 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.647092 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-rsl4r" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.656685 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.658332 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.663547 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.664336 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qwhxn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.668947 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.670689 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.675280 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lztwl" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.689437 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.690666 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.694447 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-pmng7" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.695301 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49shn\" (UniqueName: \"kubernetes.io/projected/93f6cd1f-8a49-4a2b-927a-0c4a5654891d-kube-api-access-49shn\") pod \"glance-operator-controller-manager-7bb46cd7d-wf2p9\" (UID: \"93f6cd1f-8a49-4a2b-927a-0c4a5654891d\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.695403 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6x4v\" (UniqueName: \"kubernetes.io/projected/7a378ef8-b17b-465d-aabb-e289cd9719cf-kube-api-access-g6x4v\") pod \"designate-operator-controller-manager-687df44cdb-ltlvn\" (UID: \"7a378ef8-b17b-465d-aabb-e289cd9719cf\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.695555 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztdjr\" (UniqueName: \"kubernetes.io/projected/efab1e79-db75-4c28-9046-f5fb22a49c76-kube-api-access-ztdjr\") pod \"barbican-operator-controller-manager-64f84fcdbb-srpr5\" (UID: \"efab1e79-db75-4c28-9046-f5fb22a49c76\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.695660 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lsvg\" (UniqueName: \"kubernetes.io/projected/7809e838-a0e6-4bc7-828f-bc6aaa4ebca5-kube-api-access-9lsvg\") pod \"cinder-operator-controller-manager-59cdc64769-hcblt\" (UID: \"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.706259 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.713782 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.714813 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.717513 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-2d4gl" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.742723 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.748517 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.751795 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.777694 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.778739 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.791004 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mczst" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.803087 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztdjr\" (UniqueName: \"kubernetes.io/projected/efab1e79-db75-4c28-9046-f5fb22a49c76-kube-api-access-ztdjr\") pod \"barbican-operator-controller-manager-64f84fcdbb-srpr5\" (UID: \"efab1e79-db75-4c28-9046-f5fb22a49c76\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.803144 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lsvg\" (UniqueName: \"kubernetes.io/projected/7809e838-a0e6-4bc7-828f-bc6aaa4ebca5-kube-api-access-9lsvg\") pod \"cinder-operator-controller-manager-59cdc64769-hcblt\" (UID: \"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.803180 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49shn\" (UniqueName: \"kubernetes.io/projected/93f6cd1f-8a49-4a2b-927a-0c4a5654891d-kube-api-access-49shn\") pod \"glance-operator-controller-manager-7bb46cd7d-wf2p9\" (UID: \"93f6cd1f-8a49-4a2b-927a-0c4a5654891d\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.803203 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6x4v\" (UniqueName: \"kubernetes.io/projected/7a378ef8-b17b-465d-aabb-e289cd9719cf-kube-api-access-g6x4v\") pod \"designate-operator-controller-manager-687df44cdb-ltlvn\" (UID: \"7a378ef8-b17b-465d-aabb-e289cd9719cf\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.816231 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.844266 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.844818 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lsvg\" (UniqueName: \"kubernetes.io/projected/7809e838-a0e6-4bc7-828f-bc6aaa4ebca5-kube-api-access-9lsvg\") pod \"cinder-operator-controller-manager-59cdc64769-hcblt\" (UID: \"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.845568 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.847879 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztdjr\" (UniqueName: \"kubernetes.io/projected/efab1e79-db75-4c28-9046-f5fb22a49c76-kube-api-access-ztdjr\") pod \"barbican-operator-controller-manager-64f84fcdbb-srpr5\" (UID: \"efab1e79-db75-4c28-9046-f5fb22a49c76\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.851567 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6x4v\" (UniqueName: \"kubernetes.io/projected/7a378ef8-b17b-465d-aabb-e289cd9719cf-kube-api-access-g6x4v\") pod \"designate-operator-controller-manager-687df44cdb-ltlvn\" (UID: \"7a378ef8-b17b-465d-aabb-e289cd9719cf\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.854712 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-b8cgr" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.855035 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.861626 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49shn\" (UniqueName: \"kubernetes.io/projected/93f6cd1f-8a49-4a2b-927a-0c4a5654891d-kube-api-access-49shn\") pod \"glance-operator-controller-manager-7bb46cd7d-wf2p9\" (UID: \"93f6cd1f-8a49-4a2b-927a-0c4a5654891d\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.861830 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.864018 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.867876 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-xn6bj" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.868026 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.870067 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.876408 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.877134 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-nvjkc" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.886226 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.888003 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.889127 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.897877 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tfmph" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.904292 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2ss8\" (UniqueName: \"kubernetes.io/projected/421e19af-df3c-4736-a6d1-ba43a6f8856d-kube-api-access-f2ss8\") pod \"horizon-operator-controller-manager-6d74794d9b-xg5q2\" (UID: \"421e19af-df3c-4736-a6d1-ba43a6f8856d\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.904339 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/2925bcf7-ade2-4da7-a99e-3d187eafd11a-kube-api-access-qw9jc\") pod \"heat-operator-controller-manager-6d9967f8dd-c74gx\" (UID: \"2925bcf7-ade2-4da7-a99e-3d187eafd11a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.908885 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-64wl2"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.919668 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.919779 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.924862 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b7gwk" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.930538 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.939688 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-64wl2"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.960493 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.962766 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.965110 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-vppbs" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.971990 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.974112 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.974171 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.976173 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-ltcs7" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.976298 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.978631 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.982132 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.983875 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-px8c6" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.985294 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.985647 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.991901 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6"] Oct 13 12:56:56 crc kubenswrapper[4678]: I1013 12:56:56.999000 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.000011 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.000322 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.000767 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.000885 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.002916 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.004547 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-wmvzd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.004592 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.004732 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.004947 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rj59w" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009228 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009538 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-c9npn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009651 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktnlb\" (UniqueName: \"kubernetes.io/projected/c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1-kube-api-access-ktnlb\") pod \"keystone-operator-controller-manager-ddb98f99b-qwthm\" (UID: \"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009716 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2ss8\" (UniqueName: \"kubernetes.io/projected/421e19af-df3c-4736-a6d1-ba43a6f8856d-kube-api-access-f2ss8\") pod \"horizon-operator-controller-manager-6d74794d9b-xg5q2\" (UID: \"421e19af-df3c-4736-a6d1-ba43a6f8856d\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009755 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv5sk\" (UniqueName: \"kubernetes.io/projected/11cc73f1-36f5-4221-9b67-fb4e7611ae65-kube-api-access-wv5sk\") pod \"ironic-operator-controller-manager-74cb5cbc49-vfc96\" (UID: \"11cc73f1-36f5-4221-9b67-fb4e7611ae65\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009776 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztctr\" (UniqueName: \"kubernetes.io/projected/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-kube-api-access-ztctr\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009815 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/2925bcf7-ade2-4da7-a99e-3d187eafd11a-kube-api-access-qw9jc\") pod \"heat-operator-controller-manager-6d9967f8dd-c74gx\" (UID: \"2925bcf7-ade2-4da7-a99e-3d187eafd11a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009867 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-cert\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.009930 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cwwv\" (UniqueName: \"kubernetes.io/projected/781f6537-803f-476e-a741-2030455c777f-kube-api-access-7cwwv\") pod \"mariadb-operator-controller-manager-5777b4f897-796dd\" (UID: \"781f6537-803f-476e-a741-2030455c777f\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.015647 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.019331 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.032984 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2ss8\" (UniqueName: \"kubernetes.io/projected/421e19af-df3c-4736-a6d1-ba43a6f8856d-kube-api-access-f2ss8\") pod \"horizon-operator-controller-manager-6d74794d9b-xg5q2\" (UID: \"421e19af-df3c-4736-a6d1-ba43a6f8856d\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.035528 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/2925bcf7-ade2-4da7-a99e-3d187eafd11a-kube-api-access-qw9jc\") pod \"heat-operator-controller-manager-6d9967f8dd-c74gx\" (UID: \"2925bcf7-ade2-4da7-a99e-3d187eafd11a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.035720 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.052727 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.054482 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.056208 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-f8dzc" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.071949 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.090047 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112607 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv5sk\" (UniqueName: \"kubernetes.io/projected/11cc73f1-36f5-4221-9b67-fb4e7611ae65-kube-api-access-wv5sk\") pod \"ironic-operator-controller-manager-74cb5cbc49-vfc96\" (UID: \"11cc73f1-36f5-4221-9b67-fb4e7611ae65\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112683 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztctr\" (UniqueName: \"kubernetes.io/projected/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-kube-api-access-ztctr\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112725 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gmnc\" (UniqueName: \"kubernetes.io/projected/b31e9b57-89fe-42a7-ad38-3ae06224ae58-kube-api-access-6gmnc\") pod \"octavia-operator-controller-manager-6d7c7ddf95-h9vf6\" (UID: \"b31e9b57-89fe-42a7-ad38-3ae06224ae58\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112759 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgv65\" (UniqueName: \"kubernetes.io/projected/872c1819-1912-4171-8e47-35843008daaf-kube-api-access-vgv65\") pod \"manila-operator-controller-manager-59578bc799-64wl2\" (UID: \"872c1819-1912-4171-8e47-35843008daaf\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112790 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gs4q\" (UniqueName: \"kubernetes.io/projected/d89268fa-c74a-49e5-9988-e6a9fb33d5c4-kube-api-access-7gs4q\") pod \"nova-operator-controller-manager-57bb74c7bf-hgmrn\" (UID: \"d89268fa-c74a-49e5-9988-e6a9fb33d5c4\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112826 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-cert\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112860 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lxsz\" (UniqueName: \"kubernetes.io/projected/a2deeac1-517e-4e4f-974a-1f0267868448-kube-api-access-6lxsz\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112892 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw85p\" (UniqueName: \"kubernetes.io/projected/b2063d32-c006-427e-b9a6-84007b0c1d22-kube-api-access-rw85p\") pod \"placement-operator-controller-manager-664664cb68-wvf4s\" (UID: \"b2063d32-c006-427e-b9a6-84007b0c1d22\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112924 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.112966 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cwwv\" (UniqueName: \"kubernetes.io/projected/781f6537-803f-476e-a741-2030455c777f-kube-api-access-7cwwv\") pod \"mariadb-operator-controller-manager-5777b4f897-796dd\" (UID: \"781f6537-803f-476e-a741-2030455c777f\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.113015 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g78bd\" (UniqueName: \"kubernetes.io/projected/0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a-kube-api-access-g78bd\") pod \"neutron-operator-controller-manager-797d478b46-f4f4z\" (UID: \"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.113082 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktnlb\" (UniqueName: \"kubernetes.io/projected/c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1-kube-api-access-ktnlb\") pod \"keystone-operator-controller-manager-ddb98f99b-qwthm\" (UID: \"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.113127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szw4b\" (UniqueName: \"kubernetes.io/projected/953d626f-b1c6-492b-8473-52da7c6edc3f-kube-api-access-szw4b\") pod \"ovn-operator-controller-manager-869cc7797f-5rt5t\" (UID: \"953d626f-b1c6-492b-8473-52da7c6edc3f\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.131889 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.132008 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.135307 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cwwv\" (UniqueName: \"kubernetes.io/projected/781f6537-803f-476e-a741-2030455c777f-kube-api-access-7cwwv\") pod \"mariadb-operator-controller-manager-5777b4f897-796dd\" (UID: \"781f6537-803f-476e-a741-2030455c777f\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.137387 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-56mzr" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.140636 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-cert\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.151128 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztctr\" (UniqueName: \"kubernetes.io/projected/cc1d3202-a214-4cbb-a50a-7d3ecb20409f-kube-api-access-ztctr\") pod \"infra-operator-controller-manager-585fc5b659-xb7kd\" (UID: \"cc1d3202-a214-4cbb-a50a-7d3ecb20409f\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.151875 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.162334 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktnlb\" (UniqueName: \"kubernetes.io/projected/c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1-kube-api-access-ktnlb\") pod \"keystone-operator-controller-manager-ddb98f99b-qwthm\" (UID: \"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.167370 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.170183 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.180690 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7mvvc" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.216945 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv5sk\" (UniqueName: \"kubernetes.io/projected/11cc73f1-36f5-4221-9b67-fb4e7611ae65-kube-api-access-wv5sk\") pod \"ironic-operator-controller-manager-74cb5cbc49-vfc96\" (UID: \"11cc73f1-36f5-4221-9b67-fb4e7611ae65\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217475 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wch4h\" (UniqueName: \"kubernetes.io/projected/2e600274-90bb-46a9-b41a-f642ed284900-kube-api-access-wch4h\") pod \"swift-operator-controller-manager-5f4d5dfdc6-v5b9q\" (UID: \"2e600274-90bb-46a9-b41a-f642ed284900\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217511 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szw4b\" (UniqueName: \"kubernetes.io/projected/953d626f-b1c6-492b-8473-52da7c6edc3f-kube-api-access-szw4b\") pod \"ovn-operator-controller-manager-869cc7797f-5rt5t\" (UID: \"953d626f-b1c6-492b-8473-52da7c6edc3f\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217533 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gmnc\" (UniqueName: \"kubernetes.io/projected/b31e9b57-89fe-42a7-ad38-3ae06224ae58-kube-api-access-6gmnc\") pod \"octavia-operator-controller-manager-6d7c7ddf95-h9vf6\" (UID: \"b31e9b57-89fe-42a7-ad38-3ae06224ae58\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217551 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgv65\" (UniqueName: \"kubernetes.io/projected/872c1819-1912-4171-8e47-35843008daaf-kube-api-access-vgv65\") pod \"manila-operator-controller-manager-59578bc799-64wl2\" (UID: \"872c1819-1912-4171-8e47-35843008daaf\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217570 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gs4q\" (UniqueName: \"kubernetes.io/projected/d89268fa-c74a-49e5-9988-e6a9fb33d5c4-kube-api-access-7gs4q\") pod \"nova-operator-controller-manager-57bb74c7bf-hgmrn\" (UID: \"d89268fa-c74a-49e5-9988-e6a9fb33d5c4\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217593 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw85p\" (UniqueName: \"kubernetes.io/projected/b2063d32-c006-427e-b9a6-84007b0c1d22-kube-api-access-rw85p\") pod \"placement-operator-controller-manager-664664cb68-wvf4s\" (UID: \"b2063d32-c006-427e-b9a6-84007b0c1d22\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217611 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lxsz\" (UniqueName: \"kubernetes.io/projected/a2deeac1-517e-4e4f-974a-1f0267868448-kube-api-access-6lxsz\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217632 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g78bd\" (UniqueName: \"kubernetes.io/projected/0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a-kube-api-access-g78bd\") pod \"neutron-operator-controller-manager-797d478b46-f4f4z\" (UID: \"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.217972 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:56:57 crc kubenswrapper[4678]: E1013 12:56:57.218907 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 12:56:57 crc kubenswrapper[4678]: E1013 12:56:57.218948 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert podName:a2deeac1-517e-4e4f-974a-1f0267868448 nodeName:}" failed. No retries permitted until 2025-10-13 12:56:57.718935945 +0000 UTC m=+805.803473829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" (UID: "a2deeac1-517e-4e4f-974a-1f0267868448") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.222634 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.240484 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.249358 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgv65\" (UniqueName: \"kubernetes.io/projected/872c1819-1912-4171-8e47-35843008daaf-kube-api-access-vgv65\") pod \"manila-operator-controller-manager-59578bc799-64wl2\" (UID: \"872c1819-1912-4171-8e47-35843008daaf\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.251109 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.252263 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.266859 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.270213 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw85p\" (UniqueName: \"kubernetes.io/projected/b2063d32-c006-427e-b9a6-84007b0c1d22-kube-api-access-rw85p\") pod \"placement-operator-controller-manager-664664cb68-wvf4s\" (UID: \"b2063d32-c006-427e-b9a6-84007b0c1d22\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.270776 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gmnc\" (UniqueName: \"kubernetes.io/projected/b31e9b57-89fe-42a7-ad38-3ae06224ae58-kube-api-access-6gmnc\") pod \"octavia-operator-controller-manager-6d7c7ddf95-h9vf6\" (UID: \"b31e9b57-89fe-42a7-ad38-3ae06224ae58\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.272938 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szw4b\" (UniqueName: \"kubernetes.io/projected/953d626f-b1c6-492b-8473-52da7c6edc3f-kube-api-access-szw4b\") pod \"ovn-operator-controller-manager-869cc7797f-5rt5t\" (UID: \"953d626f-b1c6-492b-8473-52da7c6edc3f\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.273577 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gs4q\" (UniqueName: \"kubernetes.io/projected/d89268fa-c74a-49e5-9988-e6a9fb33d5c4-kube-api-access-7gs4q\") pod \"nova-operator-controller-manager-57bb74c7bf-hgmrn\" (UID: \"d89268fa-c74a-49e5-9988-e6a9fb33d5c4\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.287883 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lxsz\" (UniqueName: \"kubernetes.io/projected/a2deeac1-517e-4e4f-974a-1f0267868448-kube-api-access-6lxsz\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.288576 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g78bd\" (UniqueName: \"kubernetes.io/projected/0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a-kube-api-access-g78bd\") pod \"neutron-operator-controller-manager-797d478b46-f4f4z\" (UID: \"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.327783 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wch4h\" (UniqueName: \"kubernetes.io/projected/2e600274-90bb-46a9-b41a-f642ed284900-kube-api-access-wch4h\") pod \"swift-operator-controller-manager-5f4d5dfdc6-v5b9q\" (UID: \"2e600274-90bb-46a9-b41a-f642ed284900\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.327826 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5b8h\" (UniqueName: \"kubernetes.io/projected/b7188e6b-9e98-477e-bbe8-778da7f9bda1-kube-api-access-g5b8h\") pod \"telemetry-operator-controller-manager-578874c84d-qzgc8\" (UID: \"b7188e6b-9e98-477e-bbe8-778da7f9bda1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.327848 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ljvd\" (UniqueName: \"kubernetes.io/projected/83789dd0-5367-44d6-8cf9-0f4bf92280b1-kube-api-access-4ljvd\") pod \"test-operator-controller-manager-ffcdd6c94-46shf\" (UID: \"83789dd0-5367-44d6-8cf9-0f4bf92280b1\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.335698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.344239 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.345428 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.353944 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-t89lx" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.357192 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wch4h\" (UniqueName: \"kubernetes.io/projected/2e600274-90bb-46a9-b41a-f642ed284900-kube-api-access-wch4h\") pod \"swift-operator-controller-manager-5f4d5dfdc6-v5b9q\" (UID: \"2e600274-90bb-46a9-b41a-f642ed284900\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.375362 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.388264 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.391536 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.414748 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.430913 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5b8h\" (UniqueName: \"kubernetes.io/projected/b7188e6b-9e98-477e-bbe8-778da7f9bda1-kube-api-access-g5b8h\") pod \"telemetry-operator-controller-manager-578874c84d-qzgc8\" (UID: \"b7188e6b-9e98-477e-bbe8-778da7f9bda1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.431155 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ljvd\" (UniqueName: \"kubernetes.io/projected/83789dd0-5367-44d6-8cf9-0f4bf92280b1-kube-api-access-4ljvd\") pod \"test-operator-controller-manager-ffcdd6c94-46shf\" (UID: \"83789dd0-5367-44d6-8cf9-0f4bf92280b1\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.458777 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5b8h\" (UniqueName: \"kubernetes.io/projected/b7188e6b-9e98-477e-bbe8-778da7f9bda1-kube-api-access-g5b8h\") pod \"telemetry-operator-controller-manager-578874c84d-qzgc8\" (UID: \"b7188e6b-9e98-477e-bbe8-778da7f9bda1\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.464599 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.473044 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.474511 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.478448 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-dv7qn" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.484009 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.490210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ljvd\" (UniqueName: \"kubernetes.io/projected/83789dd0-5367-44d6-8cf9-0f4bf92280b1-kube-api-access-4ljvd\") pod \"test-operator-controller-manager-ffcdd6c94-46shf\" (UID: \"83789dd0-5367-44d6-8cf9-0f4bf92280b1\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.514449 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.521216 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.534936 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.535893 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.536127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkdtn\" (UniqueName: \"kubernetes.io/projected/ec1fafb9-7b71-45b3-ade5-6df84a9fa568-kube-api-access-pkdtn\") pod \"watcher-operator-controller-manager-646675d848-8dkj8\" (UID: \"ec1fafb9-7b71-45b3-ade5-6df84a9fa568\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.538973 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8xrxv" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.540647 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.552626 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.583864 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.586430 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.598817 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.637555 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77qj2\" (UniqueName: \"kubernetes.io/projected/3137ba4b-0e52-456e-9951-08989c113034-kube-api-access-77qj2\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.637630 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkdtn\" (UniqueName: \"kubernetes.io/projected/ec1fafb9-7b71-45b3-ade5-6df84a9fa568-kube-api-access-pkdtn\") pod \"watcher-operator-controller-manager-646675d848-8dkj8\" (UID: \"ec1fafb9-7b71-45b3-ade5-6df84a9fa568\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.637752 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg7qf\" (UniqueName: \"kubernetes.io/projected/6083ce11-ee7f-4bbc-b5ff-c2816dfe6300-kube-api-access-sg7qf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz\" (UID: \"6083ce11-ee7f-4bbc-b5ff-c2816dfe6300\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.637806 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.661563 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkdtn\" (UniqueName: \"kubernetes.io/projected/ec1fafb9-7b71-45b3-ade5-6df84a9fa568-kube-api-access-pkdtn\") pod \"watcher-operator-controller-manager-646675d848-8dkj8\" (UID: \"ec1fafb9-7b71-45b3-ade5-6df84a9fa568\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.710893 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.735610 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.739527 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77qj2\" (UniqueName: \"kubernetes.io/projected/3137ba4b-0e52-456e-9951-08989c113034-kube-api-access-77qj2\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.739591 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg7qf\" (UniqueName: \"kubernetes.io/projected/6083ce11-ee7f-4bbc-b5ff-c2816dfe6300-kube-api-access-sg7qf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz\" (UID: \"6083ce11-ee7f-4bbc-b5ff-c2816dfe6300\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.739614 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.739641 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: E1013 12:56:57.739819 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 13 12:56:57 crc kubenswrapper[4678]: E1013 12:56:57.739902 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert podName:3137ba4b-0e52-456e-9951-08989c113034 nodeName:}" failed. No retries permitted until 2025-10-13 12:56:58.239878445 +0000 UTC m=+806.324416329 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert") pod "openstack-operator-controller-manager-74d8c8d84c-lsmxm" (UID: "3137ba4b-0e52-456e-9951-08989c113034") : secret "webhook-server-cert" not found Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.759798 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77qj2\" (UniqueName: \"kubernetes.io/projected/3137ba4b-0e52-456e-9951-08989c113034-kube-api-access-77qj2\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.759905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg7qf\" (UniqueName: \"kubernetes.io/projected/6083ce11-ee7f-4bbc-b5ff-c2816dfe6300-kube-api-access-sg7qf\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz\" (UID: \"6083ce11-ee7f-4bbc-b5ff-c2816dfe6300\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.766703 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a2deeac1-517e-4e4f-974a-1f0267868448-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z\" (UID: \"a2deeac1-517e-4e4f-974a-1f0267868448\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:57 crc kubenswrapper[4678]: W1013 12:56:57.771776 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93f6cd1f_8a49_4a2b_927a_0c4a5654891d.slice/crio-e4a318e481d4f50af6167a23d11a69b378610a0f2a3d836689d939c514380436 WatchSource:0}: Error finding container e4a318e481d4f50af6167a23d11a69b378610a0f2a3d836689d939c514380436: Status 404 returned error can't find the container with id e4a318e481d4f50af6167a23d11a69b378610a0f2a3d836689d939c514380436 Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.830290 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.848113 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn"] Oct 13 12:56:57 crc kubenswrapper[4678]: I1013 12:56:57.862698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.038977 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.048625 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.056146 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:56:58 crc kubenswrapper[4678]: W1013 12:56:58.110566 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11cc73f1_36f5_4221_9b67_fb4e7611ae65.slice/crio-f4879c207a8c76a621b911cddae080b40acf11e8c2439182d59a2cfb32e16190 WatchSource:0}: Error finding container f4879c207a8c76a621b911cddae080b40acf11e8c2439182d59a2cfb32e16190: Status 404 returned error can't find the container with id f4879c207a8c76a621b911cddae080b40acf11e8c2439182d59a2cfb32e16190 Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.261578 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.266123 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3137ba4b-0e52-456e-9951-08989c113034-cert\") pod \"openstack-operator-controller-manager-74d8c8d84c-lsmxm\" (UID: \"3137ba4b-0e52-456e-9951-08989c113034\") " pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.297739 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-64wl2"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.401499 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.477282 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd"] Oct 13 12:56:58 crc kubenswrapper[4678]: W1013 12:56:58.491836 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc1d3202_a214_4cbb_a50a_7d3ecb20409f.slice/crio-29f2bb25949df57fe64a620aba24104804cf6b1b5d5dfa0436eec75e5f80d4dc WatchSource:0}: Error finding container 29f2bb25949df57fe64a620aba24104804cf6b1b5d5dfa0436eec75e5f80d4dc: Status 404 returned error can't find the container with id 29f2bb25949df57fe64a620aba24104804cf6b1b5d5dfa0436eec75e5f80d4dc Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.501690 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.616431 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" event={"ID":"efab1e79-db75-4c28-9046-f5fb22a49c76","Type":"ContainerStarted","Data":"ef88a284e52e5867a3a473db3f9faadf3aa1d1a1342470af83fe002585061c53"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.617894 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" event={"ID":"872c1819-1912-4171-8e47-35843008daaf","Type":"ContainerStarted","Data":"3792dd310e1a936590f14c911683d77a93dc147fa8d61e7968daafc198b1cb09"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.620831 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" event={"ID":"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5","Type":"ContainerStarted","Data":"1c9b8e8248e32652c22d9e5b8bb5d28676873ba6ba79654a534f960d9be69da6"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.622034 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" event={"ID":"421e19af-df3c-4736-a6d1-ba43a6f8856d","Type":"ContainerStarted","Data":"5dd57b5abe8bfa46dfd6e4e430b7dee4adb3f551c2974c635bce28dc47f9f6b9"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.622846 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" event={"ID":"781f6537-803f-476e-a741-2030455c777f","Type":"ContainerStarted","Data":"e820862eb9f76d269ad101784871c3b472be86a6fa52ea15f7ea0f72f49a7b52"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.624910 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" event={"ID":"93f6cd1f-8a49-4a2b-927a-0c4a5654891d","Type":"ContainerStarted","Data":"e4a318e481d4f50af6167a23d11a69b378610a0f2a3d836689d939c514380436"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.628959 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" event={"ID":"11cc73f1-36f5-4221-9b67-fb4e7611ae65","Type":"ContainerStarted","Data":"f4879c207a8c76a621b911cddae080b40acf11e8c2439182d59a2cfb32e16190"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.629901 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" event={"ID":"cc1d3202-a214-4cbb-a50a-7d3ecb20409f","Type":"ContainerStarted","Data":"29f2bb25949df57fe64a620aba24104804cf6b1b5d5dfa0436eec75e5f80d4dc"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.634130 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" event={"ID":"7a378ef8-b17b-465d-aabb-e289cd9719cf","Type":"ContainerStarted","Data":"129b0b7eea1926f817e7a72e6550f3ea939fc2f50dc7070d4d163a927ee374e4"} Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.766313 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.769943 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.777942 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.784597 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.811241 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.817824 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.824005 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q"] Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.825135 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf"] Oct 13 12:56:58 crc kubenswrapper[4678]: W1013 12:56:58.827606 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e600274_90bb_46a9_b41a_f642ed284900.slice/crio-83d8010025009d3d5e10ee9593fd9190c59c981b32f784fcc3470324c67ad611 WatchSource:0}: Error finding container 83d8010025009d3d5e10ee9593fd9190c59c981b32f784fcc3470324c67ad611: Status 404 returned error can't find the container with id 83d8010025009d3d5e10ee9593fd9190c59c981b32f784fcc3470324c67ad611 Oct 13 12:56:58 crc kubenswrapper[4678]: W1013 12:56:58.829862 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2063d32_c006_427e_b9a6_84007b0c1d22.slice/crio-7c2e4bff39a7b340e1a3e4b4dde5e0f3f16fd35de9de9460d02d8c0e9d92a0b1 WatchSource:0}: Error finding container 7c2e4bff39a7b340e1a3e4b4dde5e0f3f16fd35de9de9460d02d8c0e9d92a0b1: Status 404 returned error can't find the container with id 7c2e4bff39a7b340e1a3e4b4dde5e0f3f16fd35de9de9460d02d8c0e9d92a0b1 Oct 13 12:56:58 crc kubenswrapper[4678]: E1013 12:56:58.834685 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wch4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-v5b9q_openstack-operators(2e600274-90bb-46a9-b41a-f642ed284900): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 12:56:58 crc kubenswrapper[4678]: W1013 12:56:58.838683 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5d4ef54_8aad_4cf4_9d10_97f3ceb91df1.slice/crio-f1cf553d05c27cc05b2d0ab878261fe92cb4e81c7789c323ea0a006b63818c13 WatchSource:0}: Error finding container f1cf553d05c27cc05b2d0ab878261fe92cb4e81c7789c323ea0a006b63818c13: Status 404 returned error can't find the container with id f1cf553d05c27cc05b2d0ab878261fe92cb4e81c7789c323ea0a006b63818c13 Oct 13 12:56:58 crc kubenswrapper[4678]: I1013 12:56:58.932479 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z"] Oct 13 12:56:58 crc kubenswrapper[4678]: E1013 12:56:58.948483 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6lxsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z_openstack-operators(a2deeac1-517e-4e4f-974a-1f0267868448): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.111378 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t"] Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.119572 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz"] Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.131102 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8"] Oct 13 12:56:59 crc kubenswrapper[4678]: W1013 12:56:59.134219 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod953d626f_b1c6_492b_8473_52da7c6edc3f.slice/crio-c916b6b576481329b92869b76ed806225092250cb362beded3b275f75d44d84f WatchSource:0}: Error finding container c916b6b576481329b92869b76ed806225092250cb362beded3b275f75d44d84f: Status 404 returned error can't find the container with id c916b6b576481329b92869b76ed806225092250cb362beded3b275f75d44d84f Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.136459 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pkdtn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-8dkj8_openstack-operators(ec1fafb9-7b71-45b3-ade5-6df84a9fa568): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.136870 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-szw4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-869cc7797f-5rt5t_openstack-operators(953d626f-b1c6-492b-8473-52da7c6edc3f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.143064 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8"] Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.144889 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g5b8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-578874c84d-qzgc8_openstack-operators(b7188e6b-9e98-477e-bbe8-778da7f9bda1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.145460 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm"] Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.327276 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" podUID="2e600274-90bb-46a9-b41a-f642ed284900" Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.340592 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" podUID="b7188e6b-9e98-477e-bbe8-778da7f9bda1" Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.351473 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" podUID="a2deeac1-517e-4e4f-974a-1f0267868448" Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.351743 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" podUID="953d626f-b1c6-492b-8473-52da7c6edc3f" Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.365387 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" podUID="ec1fafb9-7b71-45b3-ade5-6df84a9fa568" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.668101 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" event={"ID":"b2063d32-c006-427e-b9a6-84007b0c1d22","Type":"ContainerStarted","Data":"7c2e4bff39a7b340e1a3e4b4dde5e0f3f16fd35de9de9460d02d8c0e9d92a0b1"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.670872 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" event={"ID":"3137ba4b-0e52-456e-9951-08989c113034","Type":"ContainerStarted","Data":"2947dc546f5b6e58eb5a1eb1d6e67ad72a2e5955a81adf4c0e307f897bf76854"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.670924 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" event={"ID":"3137ba4b-0e52-456e-9951-08989c113034","Type":"ContainerStarted","Data":"685241457d2d128c3ef3aab6fc0d1c91a06fad86f35d0219dced23494f19e79a"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.679401 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" event={"ID":"b31e9b57-89fe-42a7-ad38-3ae06224ae58","Type":"ContainerStarted","Data":"9e248ab6d14c0c6643de42e6940ec590e9f8b46a7e33df15bbc27b730ed4ad24"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.688185 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" event={"ID":"a2deeac1-517e-4e4f-974a-1f0267868448","Type":"ContainerStarted","Data":"54194e7d46eacf61431a53d710a81567f9ccc6e6249d4ae0cb5e943a8efc3962"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.688226 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" event={"ID":"a2deeac1-517e-4e4f-974a-1f0267868448","Type":"ContainerStarted","Data":"01079dc7814bf5b396c8731d1ea14e9664a045e26ae76fdf27fa6a6456f47e37"} Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.689373 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" podUID="a2deeac1-517e-4e4f-974a-1f0267868448" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.692738 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" event={"ID":"83789dd0-5367-44d6-8cf9-0f4bf92280b1","Type":"ContainerStarted","Data":"d40838541593a3ba8166435731c87ece09de474931434fbbbb5d6c06d4af856a"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.707851 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" event={"ID":"2925bcf7-ade2-4da7-a99e-3d187eafd11a","Type":"ContainerStarted","Data":"df3f47a5fb6cf8759399bb403a9e89b4062775a68bc7405bbdcde2d48da29f4d"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.728243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" event={"ID":"d89268fa-c74a-49e5-9988-e6a9fb33d5c4","Type":"ContainerStarted","Data":"4d157b780cf67055b9bce5c94b8f12ac688e33a89f62a1714a5b231acbfdaf3e"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.739689 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" event={"ID":"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1","Type":"ContainerStarted","Data":"f1cf553d05c27cc05b2d0ab878261fe92cb4e81c7789c323ea0a006b63818c13"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.744575 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" event={"ID":"ec1fafb9-7b71-45b3-ade5-6df84a9fa568","Type":"ContainerStarted","Data":"a9c2f16cf70c9fed7265a2a29d6f6e2953a8caf7ee8aedda2f31c3d5aa2215fc"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.744612 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" event={"ID":"ec1fafb9-7b71-45b3-ade5-6df84a9fa568","Type":"ContainerStarted","Data":"7490cdfb26a30eda587040c600ca9a89d7cee1a02a50a2eb27a7910cb056e3a9"} Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.745999 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" podUID="ec1fafb9-7b71-45b3-ade5-6df84a9fa568" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.754726 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" event={"ID":"2e600274-90bb-46a9-b41a-f642ed284900","Type":"ContainerStarted","Data":"641af7852a7d28df1ea880faf29192bf9f144355cfedb917e33c3cb874168804"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.754765 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" event={"ID":"2e600274-90bb-46a9-b41a-f642ed284900","Type":"ContainerStarted","Data":"83d8010025009d3d5e10ee9593fd9190c59c981b32f784fcc3470324c67ad611"} Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.755687 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" podUID="2e600274-90bb-46a9-b41a-f642ed284900" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.760854 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" event={"ID":"953d626f-b1c6-492b-8473-52da7c6edc3f","Type":"ContainerStarted","Data":"550d26676722667a403b1e882f67be57168a13b44c448c325aab51283e55b8ef"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.760894 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" event={"ID":"953d626f-b1c6-492b-8473-52da7c6edc3f","Type":"ContainerStarted","Data":"c916b6b576481329b92869b76ed806225092250cb362beded3b275f75d44d84f"} Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.766419 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" podUID="953d626f-b1c6-492b-8473-52da7c6edc3f" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.770599 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" event={"ID":"b7188e6b-9e98-477e-bbe8-778da7f9bda1","Type":"ContainerStarted","Data":"2b8c2a05e207775095b905243706a052eed2ab724c1012b9ca8bc9838788b10a"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.770633 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" event={"ID":"b7188e6b-9e98-477e-bbe8-778da7f9bda1","Type":"ContainerStarted","Data":"d3688fd6bda9657f62024ab2c061b133039d38a5dbc8932a164edffd9d7c76a8"} Oct 13 12:56:59 crc kubenswrapper[4678]: E1013 12:56:59.777444 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" podUID="b7188e6b-9e98-477e-bbe8-778da7f9bda1" Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.779220 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" event={"ID":"6083ce11-ee7f-4bbc-b5ff-c2816dfe6300","Type":"ContainerStarted","Data":"934c0e8ca4c0810792af260e5bed595239329615e8010be367b59d4bbd01e3a0"} Oct 13 12:56:59 crc kubenswrapper[4678]: I1013 12:56:59.781713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" event={"ID":"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a","Type":"ContainerStarted","Data":"fcc979fa2f08f449abfba458b9787246b58b73fc75cc7cd47cf2bb04eaa459b3"} Oct 13 12:57:00 crc kubenswrapper[4678]: E1013 12:57:00.791501 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" podUID="a2deeac1-517e-4e4f-974a-1f0267868448" Oct 13 12:57:00 crc kubenswrapper[4678]: E1013 12:57:00.791831 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" podUID="2e600274-90bb-46a9-b41a-f642ed284900" Oct 13 12:57:00 crc kubenswrapper[4678]: E1013 12:57:00.791869 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" podUID="b7188e6b-9e98-477e-bbe8-778da7f9bda1" Oct 13 12:57:00 crc kubenswrapper[4678]: E1013 12:57:00.791896 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" podUID="953d626f-b1c6-492b-8473-52da7c6edc3f" Oct 13 12:57:00 crc kubenswrapper[4678]: E1013 12:57:00.791932 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" podUID="ec1fafb9-7b71-45b3-ade5-6df84a9fa568" Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.920873 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" event={"ID":"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a","Type":"ContainerStarted","Data":"81b1affc95e95f09d8101dfdede0c01e637d035bfa4285dfbedde31f7c382718"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.927504 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" event={"ID":"7a378ef8-b17b-465d-aabb-e289cd9719cf","Type":"ContainerStarted","Data":"bb09749b9f08fa503a17c38ceeb71fe3573e01f7b02b1dcf392fb25affc3ecde"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.929817 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" event={"ID":"872c1819-1912-4171-8e47-35843008daaf","Type":"ContainerStarted","Data":"9dcaf644e9849aeb612ab2f9026f5d730d7c1f41aee1500fa805ffaa0d8d897e"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.931847 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" event={"ID":"11cc73f1-36f5-4221-9b67-fb4e7611ae65","Type":"ContainerStarted","Data":"07a65c1bc77b600e581640552324176dcec07b486f18fb5139f6086b51b605eb"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.933800 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" event={"ID":"83789dd0-5367-44d6-8cf9-0f4bf92280b1","Type":"ContainerStarted","Data":"3f19ecee00a567cd6e959bf178f20ec9271c6b4f1c401c18c18737f0470925a0"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.934986 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" event={"ID":"efab1e79-db75-4c28-9046-f5fb22a49c76","Type":"ContainerStarted","Data":"1746ce8f390d781ba0fbde7d53b4d94041f8f0b65211a8a757f5fa1b3bf5584c"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.942064 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" event={"ID":"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5","Type":"ContainerStarted","Data":"21f247a2a29b8f6cc629e4c736e2298224ef71d4a5438a817e3dcee1a7431673"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.958729 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" event={"ID":"3137ba4b-0e52-456e-9951-08989c113034","Type":"ContainerStarted","Data":"e1e6acc980dd110e8a9db91f8b8de6c0bd7dbbf01840715e2e67056e5a5dd6f0"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.959836 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.963496 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" event={"ID":"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1","Type":"ContainerStarted","Data":"f054b877701b28f2a1240bd5aca247e607195923d903845953178f11c469bd0f"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.967292 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" event={"ID":"93f6cd1f-8a49-4a2b-927a-0c4a5654891d","Type":"ContainerStarted","Data":"9e9ca777644b209a9ee4cbcd95c821582afad0cad75505241f1d7687ee0f6064"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.970189 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.971186 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" event={"ID":"421e19af-df3c-4736-a6d1-ba43a6f8856d","Type":"ContainerStarted","Data":"e728e469161d3176958e2735286abb0fec9a459e53df7b7555b2800a13ad6422"} Oct 13 12:57:10 crc kubenswrapper[4678]: I1013 12:57:10.982126 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" event={"ID":"781f6537-803f-476e-a741-2030455c777f","Type":"ContainerStarted","Data":"e546bd9c87459cc825a5eb9e7da4525fac1555d23eff19562ce8d686ae2c3c29"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.006385 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-74d8c8d84c-lsmxm" podStartSLOduration=14.006371075 podStartE2EDuration="14.006371075s" podCreationTimestamp="2025-10-13 12:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:57:11.001409183 +0000 UTC m=+819.085947067" watchObservedRunningTime="2025-10-13 12:57:11.006371075 +0000 UTC m=+819.090908959" Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.990263 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" event={"ID":"11cc73f1-36f5-4221-9b67-fb4e7611ae65","Type":"ContainerStarted","Data":"157c0e68c4edf4655e4267320a8e1bf3771e4c32b4e07d45ef9684b17e8341e9"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.990669 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.992437 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" event={"ID":"efab1e79-db75-4c28-9046-f5fb22a49c76","Type":"ContainerStarted","Data":"8c8f1d6a96870bfedb31e9c076af9ad284cc3d1079fa9ea1cfecb88567e9b09d"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.993229 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.995425 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" event={"ID":"b2063d32-c006-427e-b9a6-84007b0c1d22","Type":"ContainerStarted","Data":"8550c1875ca33a7bbf6f4362929c824fde4c8bdd6e40cd057244612a8b7bd2c3"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.995454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" event={"ID":"b2063d32-c006-427e-b9a6-84007b0c1d22","Type":"ContainerStarted","Data":"eed9a81aafdfe11b194c756d4373aebfff979c097f13f378d49d28d4a7861daa"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.995893 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.997748 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" event={"ID":"cc1d3202-a214-4cbb-a50a-7d3ecb20409f","Type":"ContainerStarted","Data":"bd5afd517daac06fcea87252a64c558fe8fcf0f70491caf7683f613e33c342cd"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.997774 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" event={"ID":"cc1d3202-a214-4cbb-a50a-7d3ecb20409f","Type":"ContainerStarted","Data":"cfec3e3774b3052204f10045475f190474d5e78b7057c10e5132707a1e106111"} Oct 13 12:57:11 crc kubenswrapper[4678]: I1013 12:57:11.998209 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.000159 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" event={"ID":"2925bcf7-ade2-4da7-a99e-3d187eafd11a","Type":"ContainerStarted","Data":"8618f50245aa91c5f5f8dfe54c248ceb33d5349b58304652af5a2f6255dd8f1e"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.000186 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" event={"ID":"2925bcf7-ade2-4da7-a99e-3d187eafd11a","Type":"ContainerStarted","Data":"791919a62a201d9a85e7155f7c0dfe8dff8e4b3340e951cb665ab464dd2da2d5"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.000601 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.002313 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" event={"ID":"b31e9b57-89fe-42a7-ad38-3ae06224ae58","Type":"ContainerStarted","Data":"5e731d1f3bc6bb5db09b530d1d887f1a5bf21d4447c6c3f6bddcb9a4112d395a"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.002340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" event={"ID":"b31e9b57-89fe-42a7-ad38-3ae06224ae58","Type":"ContainerStarted","Data":"821ab7aafccd0c00fef7070e146c2af4da2e1413192cdd330eeb8c5fb556af07"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.002757 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.004280 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" event={"ID":"872c1819-1912-4171-8e47-35843008daaf","Type":"ContainerStarted","Data":"92511c518dfa2e9977a0f2fdf0641eac27d04014c3b20c3e2ba970864ba255c5"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.004706 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.006015 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" event={"ID":"421e19af-df3c-4736-a6d1-ba43a6f8856d","Type":"ContainerStarted","Data":"f00eb2826d97b2a5ef177f314b8cfa2835835b1ac7d663209d198c5df4221157"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.006485 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.008703 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" event={"ID":"83789dd0-5367-44d6-8cf9-0f4bf92280b1","Type":"ContainerStarted","Data":"9b5ca0fb1cb8844f0ac523898ded8576a245f6d42c9c55d6f5ad1d58c3d91a79"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.009220 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.011446 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" event={"ID":"c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1","Type":"ContainerStarted","Data":"8786a11017805a513f13a37496ac96a52c86aceb2b22f0290f192cad60e3d2e7"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.011877 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.016535 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" podStartSLOduration=6.607617031 podStartE2EDuration="16.016522069s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.116343016 +0000 UTC m=+806.200880900" lastFinishedPulling="2025-10-13 12:57:07.525248054 +0000 UTC m=+815.609785938" observedRunningTime="2025-10-13 12:57:12.015005239 +0000 UTC m=+820.099543123" watchObservedRunningTime="2025-10-13 12:57:12.016522069 +0000 UTC m=+820.101059953" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.017807 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" event={"ID":"781f6537-803f-476e-a741-2030455c777f","Type":"ContainerStarted","Data":"3a48978ea9dc47b2fa8a2914388fec17d5af2cc99baf5f05b40d875a6d474e2b"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.017892 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.020327 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" event={"ID":"7809e838-a0e6-4bc7-828f-bc6aaa4ebca5","Type":"ContainerStarted","Data":"fe1491943bf1ccec49519afdac3e1144de302c9c97dda3ef56c2ad718b70fd7a"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.020455 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.026328 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" event={"ID":"0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a","Type":"ContainerStarted","Data":"95777bf45d8fff642f9cdbeb6f4799a3d1e5f0764602491aee2f84056bab9ebf"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.026737 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.031089 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" event={"ID":"7a378ef8-b17b-465d-aabb-e289cd9719cf","Type":"ContainerStarted","Data":"3ec1a511719601604c44de1dcd549f37c194c0d38cfb9bcfae94ad80522b9036"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.031325 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.036232 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" podStartSLOduration=6.596682659 podStartE2EDuration="16.036220804s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.08572216 +0000 UTC m=+806.170260044" lastFinishedPulling="2025-10-13 12:57:07.525260305 +0000 UTC m=+815.609798189" observedRunningTime="2025-10-13 12:57:12.035461204 +0000 UTC m=+820.119999098" watchObservedRunningTime="2025-10-13 12:57:12.036220804 +0000 UTC m=+820.120758688" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.037469 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" event={"ID":"93f6cd1f-8a49-4a2b-927a-0c4a5654891d","Type":"ContainerStarted","Data":"c526eb4c805399f3afd3b558d35cfad6633e7710a4833ac575820e8dbcfaa882"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.038184 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.050019 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" event={"ID":"d89268fa-c74a-49e5-9988-e6a9fb33d5c4","Type":"ContainerStarted","Data":"7bc1a0f972822fc57c310047bb48d7bcd56b899f6877b905ff6227c6e018da36"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.050062 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" event={"ID":"d89268fa-c74a-49e5-9988-e6a9fb33d5c4","Type":"ContainerStarted","Data":"510048a6d49344f8e27068a65172719b7d58897385b111b03afec8f2dccb180b"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.050457 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.052302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" event={"ID":"6083ce11-ee7f-4bbc-b5ff-c2816dfe6300","Type":"ContainerStarted","Data":"24e20598d2131a64ce8f05edb550364ba5b39d2e4eeb9c2caa15c8838299291d"} Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.061204 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" podStartSLOduration=3.673415176 podStartE2EDuration="15.06119514s" podCreationTimestamp="2025-10-13 12:56:57 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.828092749 +0000 UTC m=+806.912630633" lastFinishedPulling="2025-10-13 12:57:10.215872723 +0000 UTC m=+818.300410597" observedRunningTime="2025-10-13 12:57:12.058336914 +0000 UTC m=+820.142874788" watchObservedRunningTime="2025-10-13 12:57:12.06119514 +0000 UTC m=+820.145733024" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.100145 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" podStartSLOduration=4.723423399 podStartE2EDuration="16.100130227s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.841028074 +0000 UTC m=+806.925565958" lastFinishedPulling="2025-10-13 12:57:10.217734902 +0000 UTC m=+818.302272786" observedRunningTime="2025-10-13 12:57:12.095195656 +0000 UTC m=+820.179733540" watchObservedRunningTime="2025-10-13 12:57:12.100130227 +0000 UTC m=+820.184668111" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.120673 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" podStartSLOduration=4.728476714 podStartE2EDuration="16.120656524s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.833324099 +0000 UTC m=+806.917861983" lastFinishedPulling="2025-10-13 12:57:10.225503899 +0000 UTC m=+818.310041793" observedRunningTime="2025-10-13 12:57:12.117533341 +0000 UTC m=+820.202071225" watchObservedRunningTime="2025-10-13 12:57:12.120656524 +0000 UTC m=+820.205194408" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.140831 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" podStartSLOduration=4.740897294 podStartE2EDuration="16.140818421s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.827322619 +0000 UTC m=+806.911860503" lastFinishedPulling="2025-10-13 12:57:10.227243746 +0000 UTC m=+818.311781630" observedRunningTime="2025-10-13 12:57:12.136025623 +0000 UTC m=+820.220563497" watchObservedRunningTime="2025-10-13 12:57:12.140818421 +0000 UTC m=+820.225356305" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.159598 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" podStartSLOduration=7.181795998 podStartE2EDuration="16.159585571s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:57.638557125 +0000 UTC m=+805.723095009" lastFinishedPulling="2025-10-13 12:57:06.616346698 +0000 UTC m=+814.700884582" observedRunningTime="2025-10-13 12:57:12.155345088 +0000 UTC m=+820.239882972" watchObservedRunningTime="2025-10-13 12:57:12.159585571 +0000 UTC m=+820.244123455" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.175462 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" podStartSLOduration=4.777755616 podStartE2EDuration="16.175446314s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.827679228 +0000 UTC m=+806.912217112" lastFinishedPulling="2025-10-13 12:57:10.225369916 +0000 UTC m=+818.309907810" observedRunningTime="2025-10-13 12:57:12.174635412 +0000 UTC m=+820.259173296" watchObservedRunningTime="2025-10-13 12:57:12.175446314 +0000 UTC m=+820.259984198" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.200600 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" podStartSLOduration=4.468316642 podStartE2EDuration="16.200581874s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.495186029 +0000 UTC m=+806.579723913" lastFinishedPulling="2025-10-13 12:57:10.227451261 +0000 UTC m=+818.311989145" observedRunningTime="2025-10-13 12:57:12.193218777 +0000 UTC m=+820.277756651" watchObservedRunningTime="2025-10-13 12:57:12.200581874 +0000 UTC m=+820.285119758" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.255711 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" podStartSLOduration=4.351814848 podStartE2EDuration="16.255695192s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.313332824 +0000 UTC m=+806.397870698" lastFinishedPulling="2025-10-13 12:57:10.217213148 +0000 UTC m=+818.301751042" observedRunningTime="2025-10-13 12:57:12.24737049 +0000 UTC m=+820.331908374" watchObservedRunningTime="2025-10-13 12:57:12.255695192 +0000 UTC m=+820.340233066" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.287381 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" podStartSLOduration=4.888833336 podStartE2EDuration="16.287365016s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.827093043 +0000 UTC m=+806.911630927" lastFinishedPulling="2025-10-13 12:57:10.225624723 +0000 UTC m=+818.310162607" observedRunningTime="2025-10-13 12:57:12.285609569 +0000 UTC m=+820.370147463" watchObservedRunningTime="2025-10-13 12:57:12.287365016 +0000 UTC m=+820.371902900" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.332290 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" podStartSLOduration=4.873638651 podStartE2EDuration="16.332258332s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.799405665 +0000 UTC m=+806.883943549" lastFinishedPulling="2025-10-13 12:57:10.258025336 +0000 UTC m=+818.342563230" observedRunningTime="2025-10-13 12:57:12.328089541 +0000 UTC m=+820.412627425" watchObservedRunningTime="2025-10-13 12:57:12.332258332 +0000 UTC m=+820.416796216" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.358838 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" podStartSLOduration=4.111789413 podStartE2EDuration="16.35881928s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:57.919930412 +0000 UTC m=+806.004468286" lastFinishedPulling="2025-10-13 12:57:10.166960249 +0000 UTC m=+818.251498153" observedRunningTime="2025-10-13 12:57:12.353872358 +0000 UTC m=+820.438410242" watchObservedRunningTime="2025-10-13 12:57:12.35881928 +0000 UTC m=+820.443357164" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.377275 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" podStartSLOduration=4.678785029 podStartE2EDuration="16.377261551s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.518673995 +0000 UTC m=+806.603211879" lastFinishedPulling="2025-10-13 12:57:10.217150517 +0000 UTC m=+818.301688401" observedRunningTime="2025-10-13 12:57:12.370802459 +0000 UTC m=+820.455340343" watchObservedRunningTime="2025-10-13 12:57:12.377261551 +0000 UTC m=+820.461799435" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.412229 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" podStartSLOduration=5.842391632 podStartE2EDuration="16.412209822s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:57.774400605 +0000 UTC m=+805.858938489" lastFinishedPulling="2025-10-13 12:57:08.344218795 +0000 UTC m=+816.428756679" observedRunningTime="2025-10-13 12:57:12.387465763 +0000 UTC m=+820.472003647" watchObservedRunningTime="2025-10-13 12:57:12.412209822 +0000 UTC m=+820.496747706" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.450237 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz" podStartSLOduration=4.360049602 podStartE2EDuration="15.450222785s" podCreationTimestamp="2025-10-13 12:56:57 +0000 UTC" firstStartedPulling="2025-10-13 12:56:59.1354885 +0000 UTC m=+807.220026384" lastFinishedPulling="2025-10-13 12:57:10.225661673 +0000 UTC m=+818.310199567" observedRunningTime="2025-10-13 12:57:12.423160974 +0000 UTC m=+820.507698848" watchObservedRunningTime="2025-10-13 12:57:12.450222785 +0000 UTC m=+820.534760669" Oct 13 12:57:12 crc kubenswrapper[4678]: I1013 12:57:12.451874 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" podStartSLOduration=4.183143733 podStartE2EDuration="16.451869449s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:57.956826055 +0000 UTC m=+806.041363929" lastFinishedPulling="2025-10-13 12:57:10.225551761 +0000 UTC m=+818.310089645" observedRunningTime="2025-10-13 12:57:12.447623926 +0000 UTC m=+820.532161800" watchObservedRunningTime="2025-10-13 12:57:12.451869449 +0000 UTC m=+820.536407333" Oct 13 12:57:15 crc kubenswrapper[4678]: I1013 12:57:15.077555 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" event={"ID":"953d626f-b1c6-492b-8473-52da7c6edc3f","Type":"ContainerStarted","Data":"b4ef5b858351ba5bf80f0fcbc54786aae18abb93893a721e83796cc329abcfc9"} Oct 13 12:57:15 crc kubenswrapper[4678]: I1013 12:57:15.078251 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:57:15 crc kubenswrapper[4678]: I1013 12:57:15.108689 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" podStartSLOduration=4.239559837 podStartE2EDuration="19.108658056s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:59.136762144 +0000 UTC m=+807.221300028" lastFinishedPulling="2025-10-13 12:57:14.005860323 +0000 UTC m=+822.090398247" observedRunningTime="2025-10-13 12:57:15.096999495 +0000 UTC m=+823.181537369" watchObservedRunningTime="2025-10-13 12:57:15.108658056 +0000 UTC m=+823.193195970" Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.091913 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" event={"ID":"2e600274-90bb-46a9-b41a-f642ed284900","Type":"ContainerStarted","Data":"c63fe5da15caf89eeb4aa3a87673a58999320a4feb12ad6c02696ab29db0cfe6"} Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.092663 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.094748 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" event={"ID":"a2deeac1-517e-4e4f-974a-1f0267868448","Type":"ContainerStarted","Data":"91e5c27061cbecc7acd87d2203665542a7795c06a764d8ec4a1d5359f0198e37"} Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.112739 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" podStartSLOduration=3.294783964 podStartE2EDuration="20.112723318s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.834279124 +0000 UTC m=+806.918817008" lastFinishedPulling="2025-10-13 12:57:15.652218478 +0000 UTC m=+823.736756362" observedRunningTime="2025-10-13 12:57:16.111754552 +0000 UTC m=+824.196292456" watchObservedRunningTime="2025-10-13 12:57:16.112723318 +0000 UTC m=+824.197261202" Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.144784 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" podStartSLOduration=3.424936652 podStartE2EDuration="20.144769162s" podCreationTimestamp="2025-10-13 12:56:56 +0000 UTC" firstStartedPulling="2025-10-13 12:56:58.948103957 +0000 UTC m=+807.032641831" lastFinishedPulling="2025-10-13 12:57:15.667936457 +0000 UTC m=+823.752474341" observedRunningTime="2025-10-13 12:57:16.141549296 +0000 UTC m=+824.226087180" watchObservedRunningTime="2025-10-13 12:57:16.144769162 +0000 UTC m=+824.229307046" Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.977963 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-srpr5" Oct 13 12:57:16 crc kubenswrapper[4678]: I1013 12:57:16.988248 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-hcblt" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.008886 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-ltlvn" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.028365 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-wf2p9" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.156091 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-xg5q2" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.222495 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vfc96" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.229558 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-xb7kd" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.246077 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-qwthm" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.257781 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-796dd" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.279916 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-64wl2" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.340747 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-c74gx" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.379614 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-hgmrn" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.401600 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-h9vf6" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.478262 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-wvf4s" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.589650 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-f4f4z" Oct 13 12:57:17 crc kubenswrapper[4678]: I1013 12:57:17.605143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-46shf" Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.056567 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.110624 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" event={"ID":"ec1fafb9-7b71-45b3-ade5-6df84a9fa568","Type":"ContainerStarted","Data":"35bc409a22556cbbb15d9695afb6ddf1d4a2b8811212f7955dd65d56e16c9052"} Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.110801 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.112546 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" event={"ID":"b7188e6b-9e98-477e-bbe8-778da7f9bda1","Type":"ContainerStarted","Data":"31174a1c2f25435847a5278977af8d76c496bff3071cebcad79f3aa48c5d260e"} Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.112711 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:57:18 crc kubenswrapper[4678]: I1013 12:57:18.128730 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" podStartSLOduration=2.448401167 podStartE2EDuration="21.128709361s" podCreationTimestamp="2025-10-13 12:56:57 +0000 UTC" firstStartedPulling="2025-10-13 12:56:59.13625724 +0000 UTC m=+807.220795124" lastFinishedPulling="2025-10-13 12:57:17.816565434 +0000 UTC m=+825.901103318" observedRunningTime="2025-10-13 12:57:18.122555357 +0000 UTC m=+826.207093241" watchObservedRunningTime="2025-10-13 12:57:18.128709361 +0000 UTC m=+826.213247245" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.385871 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" podStartSLOduration=3.739042972 podStartE2EDuration="22.385848294s" podCreationTimestamp="2025-10-13 12:56:57 +0000 UTC" firstStartedPulling="2025-10-13 12:56:59.144717215 +0000 UTC m=+807.229255099" lastFinishedPulling="2025-10-13 12:57:17.791522537 +0000 UTC m=+825.876060421" observedRunningTime="2025-10-13 12:57:18.156326007 +0000 UTC m=+826.240863891" watchObservedRunningTime="2025-10-13 12:57:19.385848294 +0000 UTC m=+827.470386178" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.390763 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.392429 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.404646 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.472241 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.472381 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csr7z\" (UniqueName: \"kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.472426 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.573825 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.573873 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.573890 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csr7z\" (UniqueName: \"kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.574403 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.574452 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.594223 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csr7z\" (UniqueName: \"kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z\") pod \"certified-operators-qfmzs\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:19 crc kubenswrapper[4678]: I1013 12:57:19.708422 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:20 crc kubenswrapper[4678]: I1013 12:57:20.185127 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:21 crc kubenswrapper[4678]: I1013 12:57:21.140671 4678 generic.go:334] "Generic (PLEG): container finished" podID="93d493db-afde-4810-8127-39b11ecb4112" containerID="b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1" exitCode=0 Oct 13 12:57:21 crc kubenswrapper[4678]: I1013 12:57:21.140726 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerDied","Data":"b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1"} Oct 13 12:57:21 crc kubenswrapper[4678]: I1013 12:57:21.141118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerStarted","Data":"9ebeefd3880cb15bfe564a7ec2738f0584a1e8a43f3a025f8a23f1aea7de54e1"} Oct 13 12:57:22 crc kubenswrapper[4678]: I1013 12:57:22.149876 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerStarted","Data":"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237"} Oct 13 12:57:23 crc kubenswrapper[4678]: I1013 12:57:23.176212 4678 generic.go:334] "Generic (PLEG): container finished" podID="93d493db-afde-4810-8127-39b11ecb4112" containerID="9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237" exitCode=0 Oct 13 12:57:23 crc kubenswrapper[4678]: I1013 12:57:23.176280 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerDied","Data":"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237"} Oct 13 12:57:24 crc kubenswrapper[4678]: I1013 12:57:24.186083 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerStarted","Data":"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65"} Oct 13 12:57:24 crc kubenswrapper[4678]: I1013 12:57:24.210628 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qfmzs" podStartSLOduration=2.786845295 podStartE2EDuration="5.210608094s" podCreationTimestamp="2025-10-13 12:57:19 +0000 UTC" firstStartedPulling="2025-10-13 12:57:21.143477354 +0000 UTC m=+829.228015278" lastFinishedPulling="2025-10-13 12:57:23.567240153 +0000 UTC m=+831.651778077" observedRunningTime="2025-10-13 12:57:24.203862424 +0000 UTC m=+832.288400318" watchObservedRunningTime="2025-10-13 12:57:24.210608094 +0000 UTC m=+832.295145988" Oct 13 12:57:27 crc kubenswrapper[4678]: I1013 12:57:27.418678 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-5rt5t" Oct 13 12:57:27 crc kubenswrapper[4678]: I1013 12:57:27.524910 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-v5b9q" Oct 13 12:57:27 crc kubenswrapper[4678]: I1013 12:57:27.546638 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-qzgc8" Oct 13 12:57:27 crc kubenswrapper[4678]: I1013 12:57:27.741997 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-8dkj8" Oct 13 12:57:28 crc kubenswrapper[4678]: I1013 12:57:28.064634 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z" Oct 13 12:57:29 crc kubenswrapper[4678]: I1013 12:57:29.709206 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:29 crc kubenswrapper[4678]: I1013 12:57:29.709662 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:29 crc kubenswrapper[4678]: I1013 12:57:29.773527 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:30 crc kubenswrapper[4678]: I1013 12:57:30.325679 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:30 crc kubenswrapper[4678]: I1013 12:57:30.385302 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.263974 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qfmzs" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="registry-server" containerID="cri-o://66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65" gracePeriod=2 Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.715629 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.798627 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities\") pod \"93d493db-afde-4810-8127-39b11ecb4112\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.798751 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csr7z\" (UniqueName: \"kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z\") pod \"93d493db-afde-4810-8127-39b11ecb4112\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.798789 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content\") pod \"93d493db-afde-4810-8127-39b11ecb4112\" (UID: \"93d493db-afde-4810-8127-39b11ecb4112\") " Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.800594 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities" (OuterVolumeSpecName: "utilities") pod "93d493db-afde-4810-8127-39b11ecb4112" (UID: "93d493db-afde-4810-8127-39b11ecb4112"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.807230 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z" (OuterVolumeSpecName: "kube-api-access-csr7z") pod "93d493db-afde-4810-8127-39b11ecb4112" (UID: "93d493db-afde-4810-8127-39b11ecb4112"). InnerVolumeSpecName "kube-api-access-csr7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.844430 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93d493db-afde-4810-8127-39b11ecb4112" (UID: "93d493db-afde-4810-8127-39b11ecb4112"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.900113 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.900166 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csr7z\" (UniqueName: \"kubernetes.io/projected/93d493db-afde-4810-8127-39b11ecb4112-kube-api-access-csr7z\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:32 crc kubenswrapper[4678]: I1013 12:57:32.900192 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93d493db-afde-4810-8127-39b11ecb4112-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.274850 4678 generic.go:334] "Generic (PLEG): container finished" podID="93d493db-afde-4810-8127-39b11ecb4112" containerID="66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65" exitCode=0 Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.274912 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerDied","Data":"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65"} Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.274925 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qfmzs" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.274948 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qfmzs" event={"ID":"93d493db-afde-4810-8127-39b11ecb4112","Type":"ContainerDied","Data":"9ebeefd3880cb15bfe564a7ec2738f0584a1e8a43f3a025f8a23f1aea7de54e1"} Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.274976 4678 scope.go:117] "RemoveContainer" containerID="66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.301411 4678 scope.go:117] "RemoveContainer" containerID="9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.312813 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.316790 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qfmzs"] Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.330137 4678 scope.go:117] "RemoveContainer" containerID="b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.392631 4678 scope.go:117] "RemoveContainer" containerID="66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65" Oct 13 12:57:33 crc kubenswrapper[4678]: E1013 12:57:33.393274 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65\": container with ID starting with 66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65 not found: ID does not exist" containerID="66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.393331 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65"} err="failed to get container status \"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65\": rpc error: code = NotFound desc = could not find container \"66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65\": container with ID starting with 66bf699fdcb50d924193f8dfecb9d9fd67f35532095ddc0fb8395748b58b7e65 not found: ID does not exist" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.393361 4678 scope.go:117] "RemoveContainer" containerID="9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237" Oct 13 12:57:33 crc kubenswrapper[4678]: E1013 12:57:33.394905 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237\": container with ID starting with 9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237 not found: ID does not exist" containerID="9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.394965 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237"} err="failed to get container status \"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237\": rpc error: code = NotFound desc = could not find container \"9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237\": container with ID starting with 9e60301d864b15ca8be5c7d4f429c8b57571b08e5a6808ab1b5f3a966744e237 not found: ID does not exist" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.395004 4678 scope.go:117] "RemoveContainer" containerID="b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1" Oct 13 12:57:33 crc kubenswrapper[4678]: E1013 12:57:33.395719 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1\": container with ID starting with b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1 not found: ID does not exist" containerID="b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1" Oct 13 12:57:33 crc kubenswrapper[4678]: I1013 12:57:33.395784 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1"} err="failed to get container status \"b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1\": rpc error: code = NotFound desc = could not find container \"b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1\": container with ID starting with b6bcf240bd356ecb7c0fe47e61af1c25842fb87f1688dca318345afd31161ee1 not found: ID does not exist" Oct 13 12:57:34 crc kubenswrapper[4678]: I1013 12:57:34.602997 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93d493db-afde-4810-8127-39b11ecb4112" path="/var/lib/kubelet/pods/93d493db-afde-4810-8127-39b11ecb4112/volumes" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.422941 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:35 crc kubenswrapper[4678]: E1013 12:57:35.423210 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="extract-content" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.423229 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="extract-content" Oct 13 12:57:35 crc kubenswrapper[4678]: E1013 12:57:35.423250 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="registry-server" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.423256 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="registry-server" Oct 13 12:57:35 crc kubenswrapper[4678]: E1013 12:57:35.423286 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="extract-utilities" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.423292 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="extract-utilities" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.423419 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d493db-afde-4810-8127-39b11ecb4112" containerName="registry-server" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.424329 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.437869 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.537739 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.538170 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.538304 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx4wb\" (UniqueName: \"kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.639703 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.639796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx4wb\" (UniqueName: \"kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.639860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.640330 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.640578 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.659977 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx4wb\" (UniqueName: \"kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb\") pod \"community-operators-xgvzg\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:35 crc kubenswrapper[4678]: I1013 12:57:35.781945 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:36 crc kubenswrapper[4678]: I1013 12:57:36.223745 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:36 crc kubenswrapper[4678]: I1013 12:57:36.295216 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerStarted","Data":"a57a0dec2661e1ddc7fad7bbdbc916b69e0faa9606efd4a7597d70ed52b8b844"} Oct 13 12:57:37 crc kubenswrapper[4678]: I1013 12:57:37.305933 4678 generic.go:334] "Generic (PLEG): container finished" podID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerID="6712987ba843bc77dbf6d567d85468ddece2958958fcc1816aed298ffa4c9415" exitCode=0 Oct 13 12:57:37 crc kubenswrapper[4678]: I1013 12:57:37.306013 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerDied","Data":"6712987ba843bc77dbf6d567d85468ddece2958958fcc1816aed298ffa4c9415"} Oct 13 12:57:38 crc kubenswrapper[4678]: I1013 12:57:38.316915 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerStarted","Data":"8334105d501f397aee3bac7c057118174cfb7f6ca7c9314b3384c4c63e4673bd"} Oct 13 12:57:39 crc kubenswrapper[4678]: I1013 12:57:39.328834 4678 generic.go:334] "Generic (PLEG): container finished" podID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerID="8334105d501f397aee3bac7c057118174cfb7f6ca7c9314b3384c4c63e4673bd" exitCode=0 Oct 13 12:57:39 crc kubenswrapper[4678]: I1013 12:57:39.329081 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerDied","Data":"8334105d501f397aee3bac7c057118174cfb7f6ca7c9314b3384c4c63e4673bd"} Oct 13 12:57:40 crc kubenswrapper[4678]: I1013 12:57:40.360426 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerStarted","Data":"3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7"} Oct 13 12:57:40 crc kubenswrapper[4678]: I1013 12:57:40.393837 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xgvzg" podStartSLOduration=2.745589639 podStartE2EDuration="5.393818238s" podCreationTimestamp="2025-10-13 12:57:35 +0000 UTC" firstStartedPulling="2025-10-13 12:57:37.308135713 +0000 UTC m=+845.392673607" lastFinishedPulling="2025-10-13 12:57:39.956364312 +0000 UTC m=+848.040902206" observedRunningTime="2025-10-13 12:57:40.392425421 +0000 UTC m=+848.476963325" watchObservedRunningTime="2025-10-13 12:57:40.393818238 +0000 UTC m=+848.478356142" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.795281 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.798513 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.800912 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7lf7m" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.803380 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.803441 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.803384 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.814216 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.861204 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n79ks\" (UniqueName: \"kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.861270 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.867308 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.875875 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.875965 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.877473 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.962107 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.962180 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftpvz\" (UniqueName: \"kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.962283 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.962418 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n79ks\" (UniqueName: \"kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.962475 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.963338 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:43 crc kubenswrapper[4678]: I1013 12:57:43.983294 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n79ks\" (UniqueName: \"kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks\") pod \"dnsmasq-dns-675f4bcbfc-xgsfr\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.063911 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.064047 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftpvz\" (UniqueName: \"kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.064130 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.064829 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.065148 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.081113 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftpvz\" (UniqueName: \"kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz\") pod \"dnsmasq-dns-78dd6ddcc-n9w64\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.122192 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.190695 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.496596 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.500641 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 12:57:44 crc kubenswrapper[4678]: I1013 12:57:44.658845 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:57:44 crc kubenswrapper[4678]: W1013 12:57:44.667121 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fc0ad8a_ab55_4760_b52a_cba1c45c3bb7.slice/crio-930b1421d9598129c5062682e6b51d0e74a4f42770b0395ce36df613382cd607 WatchSource:0}: Error finding container 930b1421d9598129c5062682e6b51d0e74a4f42770b0395ce36df613382cd607: Status 404 returned error can't find the container with id 930b1421d9598129c5062682e6b51d0e74a4f42770b0395ce36df613382cd607 Oct 13 12:57:45 crc kubenswrapper[4678]: I1013 12:57:45.398932 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" event={"ID":"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7","Type":"ContainerStarted","Data":"930b1421d9598129c5062682e6b51d0e74a4f42770b0395ce36df613382cd607"} Oct 13 12:57:45 crc kubenswrapper[4678]: I1013 12:57:45.403369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" event={"ID":"11588d48-2169-423c-b1c3-597795548526","Type":"ContainerStarted","Data":"400ff7626e73e08f3d4384b894b9e16a079ab6ffb92114283de0ec0491c88a17"} Oct 13 12:57:45 crc kubenswrapper[4678]: I1013 12:57:45.783236 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:45 crc kubenswrapper[4678]: I1013 12:57:45.783278 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:45 crc kubenswrapper[4678]: I1013 12:57:45.844011 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.253710 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.270456 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.274817 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.277256 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.400292 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.401200 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm455\" (UniqueName: \"kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.401260 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.486937 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.502963 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.502998 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm455\" (UniqueName: \"kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.503033 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.504335 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.504453 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.541235 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm455\" (UniqueName: \"kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455\") pod \"dnsmasq-dns-666b6646f7-dx9js\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.564725 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.574840 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.593327 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.612206 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.613406 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.613483 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.705249 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch9wg\" (UniqueName: \"kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.705320 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.706995 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.820575 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch9wg\" (UniqueName: \"kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.820896 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.820998 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.822373 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.822493 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.870954 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch9wg\" (UniqueName: \"kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg\") pod \"dnsmasq-dns-57d769cc4f-z2cl4\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.900090 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:57:46 crc kubenswrapper[4678]: I1013 12:57:46.983138 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.401306 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:57:47 crc kubenswrapper[4678]: W1013 12:57:47.410631 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod390d51cb_5775_4cc2_85c9_a2cd23cb9144.slice/crio-3620a7996f7785142af97805a68e509543a5ab21761ad7e7eb1406c6e0c265dd WatchSource:0}: Error finding container 3620a7996f7785142af97805a68e509543a5ab21761ad7e7eb1406c6e0c265dd: Status 404 returned error can't find the container with id 3620a7996f7785142af97805a68e509543a5ab21761ad7e7eb1406c6e0c265dd Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.428306 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" event={"ID":"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c","Type":"ContainerStarted","Data":"44b6232f0209340c34f740b6015d48f29b4786e67b91a793715fa7ce6b18e6fd"} Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.429975 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" event={"ID":"390d51cb-5775-4cc2-85c9-a2cd23cb9144","Type":"ContainerStarted","Data":"3620a7996f7785142af97805a68e509543a5ab21761ad7e7eb1406c6e0c265dd"} Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.431078 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.432758 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.438101 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.438448 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.438621 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.438733 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlgsl" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.439253 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.439914 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.439131 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.442871 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.543964 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45gf8\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544005 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544035 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544081 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544101 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544145 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544163 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544222 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544238 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.544253 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645810 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45gf8\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645887 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645915 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645937 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645975 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.645994 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.646032 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.646073 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.646090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.646106 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.647380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.647961 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.648490 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.658763 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.665917 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.667538 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.667675 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.670006 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.672386 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.677507 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45gf8\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.678477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.690198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.721725 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.723150 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.725643 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.725693 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.725765 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.725776 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.725819 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zxwsg" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.726004 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.727143 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.728723 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.763948 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848193 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848451 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848478 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848503 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848522 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848540 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5nw\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848652 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848687 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848708 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848723 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.848742 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950532 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950808 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950835 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950895 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950914 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950938 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950962 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.950983 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.951002 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5nw\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.951021 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.951044 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.951624 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.951824 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.952553 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.953088 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.953442 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.954818 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.957628 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.960113 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.966699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.970991 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.975841 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5nw\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:47 crc kubenswrapper[4678]: I1013 12:57:47.983990 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:48 crc kubenswrapper[4678]: I1013 12:57:48.088584 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:57:48 crc kubenswrapper[4678]: I1013 12:57:48.229394 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 12:57:48 crc kubenswrapper[4678]: I1013 12:57:48.436197 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xgvzg" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="registry-server" containerID="cri-o://3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" gracePeriod=2 Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.181454 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.192232 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.192339 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.195979 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.196192 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4kvkc" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.196243 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.196769 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.197295 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.213335 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276510 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276584 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276604 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276654 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkwdh\" (UniqueName: \"kubernetes.io/projected/16bc2692-fbba-4e51-92b4-805141199ee2-kube-api-access-dkwdh\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276682 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-secrets\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276724 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-kolla-config\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.276996 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.277023 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.277285 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-default\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378750 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-kolla-config\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378803 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378891 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-default\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378941 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378972 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.378993 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.379024 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkwdh\" (UniqueName: \"kubernetes.io/projected/16bc2692-fbba-4e51-92b4-805141199ee2-kube-api-access-dkwdh\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.379079 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-secrets\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.380531 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.381483 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-kolla-config\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.381648 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.384880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.389133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/16bc2692-fbba-4e51-92b4-805141199ee2-config-data-default\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.395172 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.401319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.406816 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/16bc2692-fbba-4e51-92b4-805141199ee2-secrets\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.408913 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkwdh\" (UniqueName: \"kubernetes.io/projected/16bc2692-fbba-4e51-92b4-805141199ee2-kube-api-access-dkwdh\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.433259 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-galera-0\" (UID: \"16bc2692-fbba-4e51-92b4-805141199ee2\") " pod="openstack/openstack-galera-0" Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.448820 4678 generic.go:334] "Generic (PLEG): container finished" podID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerID="3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" exitCode=0 Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.448862 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerDied","Data":"3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7"} Oct 13 12:57:49 crc kubenswrapper[4678]: I1013 12:57:49.509676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.573649 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.575391 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.582240 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.582462 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.582899 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wzqxc" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.588230 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.588653 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697668 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697762 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697806 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697854 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc8nl\" (UniqueName: \"kubernetes.io/projected/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kube-api-access-xc8nl\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697935 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697960 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.697989 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.698028 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.698187 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.799848 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc8nl\" (UniqueName: \"kubernetes.io/projected/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kube-api-access-xc8nl\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.799902 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.799929 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.799961 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800000 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800034 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800108 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800158 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800196 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.800640 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.802580 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.802928 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.803198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.803441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.807574 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.808192 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.810047 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.826160 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc8nl\" (UniqueName: \"kubernetes.io/projected/7950a23a-e8d5-4d7e-a4d5-a30f535624b3-kube-api-access-xc8nl\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.843431 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7950a23a-e8d5-4d7e-a4d5-a30f535624b3\") " pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:50 crc kubenswrapper[4678]: I1013 12:57:50.901605 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.017305 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.018355 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.021170 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5zz2v" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.021401 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.022490 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.028820 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.105012 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-kolla-config\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.105122 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-memcached-tls-certs\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.105154 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-combined-ca-bundle\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.105208 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kchb\" (UniqueName: \"kubernetes.io/projected/911b721e-cee9-45db-ae01-fb84bb431847-kube-api-access-7kchb\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.105502 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-config-data\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.207260 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-config-data\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.207303 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-kolla-config\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.207351 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-memcached-tls-certs\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.207372 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-combined-ca-bundle\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.207403 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kchb\" (UniqueName: \"kubernetes.io/projected/911b721e-cee9-45db-ae01-fb84bb431847-kube-api-access-7kchb\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.208393 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-kolla-config\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.208468 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/911b721e-cee9-45db-ae01-fb84bb431847-config-data\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.211657 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-combined-ca-bundle\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.212241 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/911b721e-cee9-45db-ae01-fb84bb431847-memcached-tls-certs\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.225585 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kchb\" (UniqueName: \"kubernetes.io/projected/911b721e-cee9-45db-ae01-fb84bb431847-kube-api-access-7kchb\") pod \"memcached-0\" (UID: \"911b721e-cee9-45db-ae01-fb84bb431847\") " pod="openstack/memcached-0" Oct 13 12:57:51 crc kubenswrapper[4678]: I1013 12:57:51.341652 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 13 12:57:52 crc kubenswrapper[4678]: W1013 12:57:52.744320 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0795d10_95f2_4bf0_b15c_863ed19b32cd.slice/crio-84fc16241189c25a6307c4d3e3089de91abcf1b3066c93eeaaa9c2a84769ac11 WatchSource:0}: Error finding container 84fc16241189c25a6307c4d3e3089de91abcf1b3066c93eeaaa9c2a84769ac11: Status 404 returned error can't find the container with id 84fc16241189c25a6307c4d3e3089de91abcf1b3066c93eeaaa9c2a84769ac11 Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.023383 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.024701 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.026569 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-55pp7" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.034622 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.137831 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg6gw\" (UniqueName: \"kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw\") pod \"kube-state-metrics-0\" (UID: \"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec\") " pod="openstack/kube-state-metrics-0" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.238817 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg6gw\" (UniqueName: \"kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw\") pod \"kube-state-metrics-0\" (UID: \"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec\") " pod="openstack/kube-state-metrics-0" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.255753 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg6gw\" (UniqueName: \"kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw\") pod \"kube-state-metrics-0\" (UID: \"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec\") " pod="openstack/kube-state-metrics-0" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.348871 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 12:57:53 crc kubenswrapper[4678]: I1013 12:57:53.479942 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerStarted","Data":"84fc16241189c25a6307c4d3e3089de91abcf1b3066c93eeaaa9c2a84769ac11"} Oct 13 12:57:55 crc kubenswrapper[4678]: I1013 12:57:55.506262 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:57:55 crc kubenswrapper[4678]: I1013 12:57:55.506326 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:57:55 crc kubenswrapper[4678]: E1013 12:57:55.784476 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7 is running failed: container process not found" containerID="3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 12:57:55 crc kubenswrapper[4678]: E1013 12:57:55.785243 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7 is running failed: container process not found" containerID="3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 12:57:55 crc kubenswrapper[4678]: E1013 12:57:55.786107 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7 is running failed: container process not found" containerID="3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" cmd=["grpc_health_probe","-addr=:50051"] Oct 13 12:57:55 crc kubenswrapper[4678]: E1013 12:57:55.786148 4678 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-xgvzg" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="registry-server" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.265478 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dv2nq"] Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.266677 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.269002 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-hjgnw" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.269859 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.270097 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.279447 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-4z6wg"] Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.296954 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.323162 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq"] Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.330080 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4z6wg"] Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411424 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgbh\" (UniqueName: \"kubernetes.io/projected/0daece72-e6c8-427f-a970-496545b49160-kube-api-access-flgbh\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411470 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-run\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411490 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-combined-ca-bundle\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411518 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-scripts\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411535 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-log\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411560 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5dtz\" (UniqueName: \"kubernetes.io/projected/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-kube-api-access-b5dtz\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411738 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411787 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-log-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411876 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-etc-ovs\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.411941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-ovn-controller-tls-certs\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.412014 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.412107 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0daece72-e6c8-427f-a970-496545b49160-scripts\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.412139 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-lib\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513173 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-etc-ovs\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513221 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-ovn-controller-tls-certs\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513256 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513290 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0daece72-e6c8-427f-a970-496545b49160-scripts\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-lib\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513328 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flgbh\" (UniqueName: \"kubernetes.io/projected/0daece72-e6c8-427f-a970-496545b49160-kube-api-access-flgbh\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513342 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-run\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513377 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-combined-ca-bundle\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513398 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-scripts\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513416 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-log\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513440 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5dtz\" (UniqueName: \"kubernetes.io/projected/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-kube-api-access-b5dtz\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513462 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513483 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-log-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513910 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-lib\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.513969 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-log-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.514065 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-etc-ovs\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.514339 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-run\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.514781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run-ovn\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.514854 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-var-run\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.515130 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0daece72-e6c8-427f-a970-496545b49160-var-log\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.516550 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-scripts\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.517708 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0daece72-e6c8-427f-a970-496545b49160-scripts\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.535742 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-ovn-controller-tls-certs\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.535866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-combined-ca-bundle\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.541409 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5dtz\" (UniqueName: \"kubernetes.io/projected/2e724ee3-a6c7-4b75-9c60-3fe2f233b96b-kube-api-access-b5dtz\") pod \"ovn-controller-dv2nq\" (UID: \"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b\") " pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.541932 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgbh\" (UniqueName: \"kubernetes.io/projected/0daece72-e6c8-427f-a970-496545b49160-kube-api-access-flgbh\") pod \"ovn-controller-ovs-4z6wg\" (UID: \"0daece72-e6c8-427f-a970-496545b49160\") " pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.596406 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq" Oct 13 12:57:56 crc kubenswrapper[4678]: I1013 12:57:56.625737 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.022201 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.135453 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx4wb\" (UniqueName: \"kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb\") pod \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.135533 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content\") pod \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.138203 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities\") pod \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\" (UID: \"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba\") " Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.139966 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities" (OuterVolumeSpecName: "utilities") pod "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" (UID: "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.141174 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb" (OuterVolumeSpecName: "kube-api-access-gx4wb") pod "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" (UID: "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba"). InnerVolumeSpecName "kube-api-access-gx4wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.170526 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 12:57:57 crc kubenswrapper[4678]: E1013 12:57:57.170867 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="registry-server" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.170880 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="registry-server" Oct 13 12:57:57 crc kubenswrapper[4678]: E1013 12:57:57.170906 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="extract-content" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.170912 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="extract-content" Oct 13 12:57:57 crc kubenswrapper[4678]: E1013 12:57:57.170931 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="extract-utilities" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.170936 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="extract-utilities" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.171110 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" containerName="registry-server" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.171918 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.174345 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.174449 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.174516 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.174518 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-bhfdt" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.174639 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.179900 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.207004 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" (UID: "52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.239997 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.240024 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.240035 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx4wb\" (UniqueName: \"kubernetes.io/projected/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba-kube-api-access-gx4wb\") on node \"crc\" DevicePath \"\"" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341416 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341466 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341493 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-config\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341520 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341554 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341578 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341785 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbsj6\" (UniqueName: \"kubernetes.io/projected/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-kube-api-access-rbsj6\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.341834 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.443637 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbsj6\" (UniqueName: \"kubernetes.io/projected/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-kube-api-access-rbsj6\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.443677 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.443723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.443743 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444094 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444118 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444128 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-config\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444272 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444642 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444600 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-config\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.444704 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.449951 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.450579 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.451021 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.462648 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbsj6\" (UniqueName: \"kubernetes.io/projected/b1da318b-1924-4ce5-9a9a-a1b483c0ffb0-kube-api-access-rbsj6\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.464695 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0\") " pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.496399 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.511469 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgvzg" event={"ID":"52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba","Type":"ContainerDied","Data":"a57a0dec2661e1ddc7fad7bbdbc916b69e0faa9606efd4a7597d70ed52b8b844"} Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.511509 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgvzg" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.511541 4678 scope.go:117] "RemoveContainer" containerID="3b4d34c0f6d421eb3b2f03120ac0cb68d0dc40d6a59d26c92a3fb8cd7e0985d7" Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.597958 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:57 crc kubenswrapper[4678]: I1013 12:57:57.605246 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xgvzg"] Oct 13 12:57:58 crc kubenswrapper[4678]: I1013 12:57:58.601314 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba" path="/var/lib/kubelet/pods/52e68d87-a2c0-4f3a-976a-77b2ccd1e5ba/volumes" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.705709 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.711451 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.713177 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.713226 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.713181 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.713977 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-pz965" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.721669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816740 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs4hl\" (UniqueName: \"kubernetes.io/projected/352ce882-938a-4fe3-b4ce-7e01a5a750e6-kube-api-access-xs4hl\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816814 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816852 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816868 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816896 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816914 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816940 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-config\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.816960 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.918861 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.918968 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919002 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919088 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919126 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919176 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-config\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919215 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919291 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs4hl\" (UniqueName: \"kubernetes.io/projected/352ce882-938a-4fe3-b4ce-7e01a5a750e6-kube-api-access-xs4hl\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919589 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.919638 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.920494 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-config\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.920881 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/352ce882-938a-4fe3-b4ce-7e01a5a750e6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.924789 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.926006 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.930849 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/352ce882-938a-4fe3-b4ce-7e01a5a750e6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.936268 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs4hl\" (UniqueName: \"kubernetes.io/projected/352ce882-938a-4fe3-b4ce-7e01a5a750e6-kube-api-access-xs4hl\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:00 crc kubenswrapper[4678]: I1013 12:58:00.943192 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"352ce882-938a-4fe3-b4ce-7e01a5a750e6\") " pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:01 crc kubenswrapper[4678]: I1013 12:58:01.043134 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.264332 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.265027 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftpvz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-n9w64_openstack(2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.266560 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" podUID="2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7" Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.327764 4678 scope.go:117] "RemoveContainer" containerID="8334105d501f397aee3bac7c057118174cfb7f6ca7c9314b3384c4c63e4673bd" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.364188 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.364412 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rm455,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-dx9js_openstack(61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.366067 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.390826 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.390972 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n79ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-xgsfr_openstack(11588d48-2169-423c-b1c3-597795548526): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:58:03 crc kubenswrapper[4678]: E1013 12:58:03.392229 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" podUID="11588d48-2169-423c-b1c3-597795548526" Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.585244 4678 scope.go:117] "RemoveContainer" containerID="6712987ba843bc77dbf6d567d85468ddece2958958fcc1816aed298ffa4c9415" Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.718758 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 13 12:58:03 crc kubenswrapper[4678]: W1013 12:58:03.740887 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16bc2692_fbba_4e51_92b4_805141199ee2.slice/crio-b4d454fd40afe4adfdc13fba88c1434c04269ca720cb30ba4c89bd91f454f2f1 WatchSource:0}: Error finding container b4d454fd40afe4adfdc13fba88c1434c04269ca720cb30ba4c89bd91f454f2f1: Status 404 returned error can't find the container with id b4d454fd40afe4adfdc13fba88c1434c04269ca720cb30ba4c89bd91f454f2f1 Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.932709 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 13 12:58:03 crc kubenswrapper[4678]: W1013 12:58:03.942912 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod911b721e_cee9_45db_ae01_fb84bb431847.slice/crio-db87ec1aa5dec7c88461d6ba6068ca0c1690e97c25216902234e173ccf4d6b35 WatchSource:0}: Error finding container db87ec1aa5dec7c88461d6ba6068ca0c1690e97c25216902234e173ccf4d6b35: Status 404 returned error can't find the container with id db87ec1aa5dec7c88461d6ba6068ca0c1690e97c25216902234e173ccf4d6b35 Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.948358 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 12:58:03 crc kubenswrapper[4678]: W1013 12:58:03.949617 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7950a23a_e8d5_4d7e_a4d5_a30f535624b3.slice/crio-ad7b5065dfe08a26a3d162f730f209eac46e2c5558d0365e8a5bcb2983ea9631 WatchSource:0}: Error finding container ad7b5065dfe08a26a3d162f730f209eac46e2c5558d0365e8a5bcb2983ea9631: Status 404 returned error can't find the container with id ad7b5065dfe08a26a3d162f730f209eac46e2c5558d0365e8a5bcb2983ea9631 Oct 13 12:58:03 crc kubenswrapper[4678]: I1013 12:58:03.952712 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.197032 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.226254 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.306019 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftpvz\" (UniqueName: \"kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz\") pod \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307084 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n79ks\" (UniqueName: \"kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks\") pod \"11588d48-2169-423c-b1c3-597795548526\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307133 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config\") pod \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307154 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config\") pod \"11588d48-2169-423c-b1c3-597795548526\" (UID: \"11588d48-2169-423c-b1c3-597795548526\") " Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307189 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc\") pod \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\" (UID: \"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7\") " Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307826 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config" (OuterVolumeSpecName: "config") pod "11588d48-2169-423c-b1c3-597795548526" (UID: "11588d48-2169-423c-b1c3-597795548526"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307846 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config" (OuterVolumeSpecName: "config") pod "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7" (UID: "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.307920 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7" (UID: "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.361620 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq"] Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.368304 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.395534 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz" (OuterVolumeSpecName: "kube-api-access-ftpvz") pod "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7" (UID: "2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7"). InnerVolumeSpecName "kube-api-access-ftpvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.396240 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks" (OuterVolumeSpecName: "kube-api-access-n79ks") pod "11588d48-2169-423c-b1c3-597795548526" (UID: "11588d48-2169-423c-b1c3-597795548526"). InnerVolumeSpecName "kube-api-access-n79ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:04 crc kubenswrapper[4678]: W1013 12:58:04.401143 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ac0428d_3902_4f97_9ca3_b8bb81ccfeec.slice/crio-37d36584c35e4061352d83f0e876c469ca98e2b50ee468d917d4ef8e88d29bef WatchSource:0}: Error finding container 37d36584c35e4061352d83f0e876c469ca98e2b50ee468d917d4ef8e88d29bef: Status 404 returned error can't find the container with id 37d36584c35e4061352d83f0e876c469ca98e2b50ee468d917d4ef8e88d29bef Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.408957 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.408991 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftpvz\" (UniqueName: \"kubernetes.io/projected/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-kube-api-access-ftpvz\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.409008 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n79ks\" (UniqueName: \"kubernetes.io/projected/11588d48-2169-423c-b1c3-597795548526-kube-api-access-n79ks\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.409021 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.409033 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11588d48-2169-423c-b1c3-597795548526-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.475206 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 13 12:58:04 crc kubenswrapper[4678]: W1013 12:58:04.498274 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e724ee3_a6c7_4b75_9c60_3fe2f233b96b.slice/crio-49fe20e45a9d157846ed423758d979b9205f6893b45fcc61a6317f27d0d234c8 WatchSource:0}: Error finding container 49fe20e45a9d157846ed423758d979b9205f6893b45fcc61a6317f27d0d234c8: Status 404 returned error can't find the container with id 49fe20e45a9d157846ed423758d979b9205f6893b45fcc61a6317f27d0d234c8 Oct 13 12:58:04 crc kubenswrapper[4678]: W1013 12:58:04.500139 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1da318b_1924_4ce5_9a9a_a1b483c0ffb0.slice/crio-3fa5a268c91c50ddac417ffdc8056ba74209271df78cb3ab0fc4f894f73638cb WatchSource:0}: Error finding container 3fa5a268c91c50ddac417ffdc8056ba74209271df78cb3ab0fc4f894f73638cb: Status 404 returned error can't find the container with id 3fa5a268c91c50ddac417ffdc8056ba74209271df78cb3ab0fc4f894f73638cb Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec","Type":"ContainerStarted","Data":"37d36584c35e4061352d83f0e876c469ca98e2b50ee468d917d4ef8e88d29bef"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"16bc2692-fbba-4e51-92b4-805141199ee2","Type":"ContainerStarted","Data":"b4d454fd40afe4adfdc13fba88c1434c04269ca720cb30ba4c89bd91f454f2f1"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608731 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0","Type":"ContainerStarted","Data":"3fa5a268c91c50ddac417ffdc8056ba74209271df78cb3ab0fc4f894f73638cb"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerStarted","Data":"23ed0708098bd2dc31620ca16f3efa29d528b78602a995a6e3a94f93a2cd5692"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608759 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"911b721e-cee9-45db-ae01-fb84bb431847","Type":"ContainerStarted","Data":"db87ec1aa5dec7c88461d6ba6068ca0c1690e97c25216902234e173ccf4d6b35"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.608771 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq" event={"ID":"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b","Type":"ContainerStarted","Data":"49fe20e45a9d157846ed423758d979b9205f6893b45fcc61a6317f27d0d234c8"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.612491 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" event={"ID":"2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7","Type":"ContainerDied","Data":"930b1421d9598129c5062682e6b51d0e74a4f42770b0395ce36df613382cd607"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.612578 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-n9w64" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.615611 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.615625 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" event={"ID":"11588d48-2169-423c-b1c3-597795548526","Type":"ContainerDied","Data":"400ff7626e73e08f3d4384b894b9e16a079ab6ffb92114283de0ec0491c88a17"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.616672 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7950a23a-e8d5-4d7e-a4d5-a30f535624b3","Type":"ContainerStarted","Data":"ad7b5065dfe08a26a3d162f730f209eac46e2c5558d0365e8a5bcb2983ea9631"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.622016 4678 generic.go:334] "Generic (PLEG): container finished" podID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerID="271f9e7dea4d7cd9e78c72888fb67d9ceb9defdd718460124052fb23efedfd9b" exitCode=0 Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.622107 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" event={"ID":"390d51cb-5775-4cc2-85c9-a2cd23cb9144","Type":"ContainerDied","Data":"271f9e7dea4d7cd9e78c72888fb67d9ceb9defdd718460124052fb23efedfd9b"} Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.624736 4678 generic.go:334] "Generic (PLEG): container finished" podID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerID="ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af" exitCode=0 Oct 13 12:58:04 crc kubenswrapper[4678]: I1013 12:58:04.624782 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" event={"ID":"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c","Type":"ContainerDied","Data":"ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af"} Oct 13 12:58:04 crc kubenswrapper[4678]: E1013 12:58:04.838629 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fc0ad8a_ab55_4760_b52a_cba1c45c3bb7.slice\": RecentStats: unable to find data in memory cache]" Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.090676 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.099510 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-n9w64"] Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.206156 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4z6wg"] Oct 13 12:58:05 crc kubenswrapper[4678]: W1013 12:58:05.236794 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0daece72_e6c8_427f_a970_496545b49160.slice/crio-02c93160a1a0c26115239e8e9c9be75a5c6cb47c636c7a58be179135b3e83e88 WatchSource:0}: Error finding container 02c93160a1a0c26115239e8e9c9be75a5c6cb47c636c7a58be179135b3e83e88: Status 404 returned error can't find the container with id 02c93160a1a0c26115239e8e9c9be75a5c6cb47c636c7a58be179135b3e83e88 Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.306735 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 13 12:58:05 crc kubenswrapper[4678]: W1013 12:58:05.511745 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod352ce882_938a_4fe3_b4ce_7e01a5a750e6.slice/crio-e57e8b7f7f74ea908fbfb70fc264ea2532b96fbdd8b45a9daded462dae2f11eb WatchSource:0}: Error finding container e57e8b7f7f74ea908fbfb70fc264ea2532b96fbdd8b45a9daded462dae2f11eb: Status 404 returned error can't find the container with id e57e8b7f7f74ea908fbfb70fc264ea2532b96fbdd8b45a9daded462dae2f11eb Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.635322 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerStarted","Data":"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.637933 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"352ce882-938a-4fe3-b4ce-7e01a5a750e6","Type":"ContainerStarted","Data":"e57e8b7f7f74ea908fbfb70fc264ea2532b96fbdd8b45a9daded462dae2f11eb"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.641525 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerStarted","Data":"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.643108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4z6wg" event={"ID":"0daece72-e6c8-427f-a970-496545b49160","Type":"ContainerStarted","Data":"02c93160a1a0c26115239e8e9c9be75a5c6cb47c636c7a58be179135b3e83e88"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.645267 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" event={"ID":"390d51cb-5775-4cc2-85c9-a2cd23cb9144","Type":"ContainerStarted","Data":"6373bdec25a6156f7e78841351fde5860bc32a8ccfe52f76e7bdaba502a33466"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.645813 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.647662 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" event={"ID":"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c","Type":"ContainerStarted","Data":"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567"} Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.648188 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.676710 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" podStartSLOduration=-9223372017.17808 podStartE2EDuration="19.67669669s" podCreationTimestamp="2025-10-13 12:57:46 +0000 UTC" firstStartedPulling="2025-10-13 12:57:46.913223869 +0000 UTC m=+854.997761753" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:05.671595054 +0000 UTC m=+873.756132938" watchObservedRunningTime="2025-10-13 12:58:05.67669669 +0000 UTC m=+873.761234574" Oct 13 12:58:05 crc kubenswrapper[4678]: I1013 12:58:05.709806 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" podStartSLOduration=3.695703385 podStartE2EDuration="19.709789522s" podCreationTimestamp="2025-10-13 12:57:46 +0000 UTC" firstStartedPulling="2025-10-13 12:57:47.412308137 +0000 UTC m=+855.496846021" lastFinishedPulling="2025-10-13 12:58:03.426394264 +0000 UTC m=+871.510932158" observedRunningTime="2025-10-13 12:58:05.708531818 +0000 UTC m=+873.793069702" watchObservedRunningTime="2025-10-13 12:58:05.709789522 +0000 UTC m=+873.794327406" Oct 13 12:58:06 crc kubenswrapper[4678]: I1013 12:58:06.601495 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7" path="/var/lib/kubelet/pods/2fc0ad8a-ab55-4760-b52a-cba1c45c3bb7/volumes" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.594247 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.714409 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq" event={"ID":"2e724ee3-a6c7-4b75-9c60-3fe2f233b96b","Type":"ContainerStarted","Data":"3cf60475830488b47f54734ff1a996b6a1fa968cbd0e1a0a5cef845391fb5416"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.714540 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-dv2nq" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.715941 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec","Type":"ContainerStarted","Data":"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.716172 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.718100 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7950a23a-e8d5-4d7e-a4d5-a30f535624b3","Type":"ContainerStarted","Data":"6565a13408225bff0c227367c8aaa374637ffcf6f06dce3803a9c3030714717a"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.719860 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"16bc2692-fbba-4e51-92b4-805141199ee2","Type":"ContainerStarted","Data":"a42e233121f90f03b6e714c49c682def1bea1cbcaa396d6f70cfef98371bc8a5"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.721564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0","Type":"ContainerStarted","Data":"23ce017451351b906c57ab0865ec8fb50b227a790b33d8925fb7ac7c14ec7972"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.722738 4678 generic.go:334] "Generic (PLEG): container finished" podID="0daece72-e6c8-427f-a970-496545b49160" containerID="ebd432884ecbed37231752a4bfcd8ae175b305f300b18bce8452d8c0045051ef" exitCode=0 Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.722786 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4z6wg" event={"ID":"0daece72-e6c8-427f-a970-496545b49160","Type":"ContainerDied","Data":"ebd432884ecbed37231752a4bfcd8ae175b305f300b18bce8452d8c0045051ef"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.725139 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"911b721e-cee9-45db-ae01-fb84bb431847","Type":"ContainerStarted","Data":"deab6f4d80cbf1ccf1ef71ae926402f8e89766f6bfbdba3e7e7d8503ea5eac40"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.725515 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.728933 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"352ce882-938a-4fe3-b4ce-7e01a5a750e6","Type":"ContainerStarted","Data":"3fbfbc3109bda212bae24e4e8208eafbbd8a307eb9d7177c0a8622f367980e14"} Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.737808 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dv2nq" podStartSLOduration=9.905792783 podStartE2EDuration="15.73778996s" podCreationTimestamp="2025-10-13 12:57:56 +0000 UTC" firstStartedPulling="2025-10-13 12:58:04.500949993 +0000 UTC m=+872.585487867" lastFinishedPulling="2025-10-13 12:58:10.33294715 +0000 UTC m=+878.417485044" observedRunningTime="2025-10-13 12:58:11.735950871 +0000 UTC m=+879.820488775" watchObservedRunningTime="2025-10-13 12:58:11.73778996 +0000 UTC m=+879.822327854" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.827003 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.210028496 podStartE2EDuration="19.826983816s" podCreationTimestamp="2025-10-13 12:57:52 +0000 UTC" firstStartedPulling="2025-10-13 12:58:04.403495627 +0000 UTC m=+872.488033521" lastFinishedPulling="2025-10-13 12:58:11.020450957 +0000 UTC m=+879.104988841" observedRunningTime="2025-10-13 12:58:11.821540941 +0000 UTC m=+879.906078825" watchObservedRunningTime="2025-10-13 12:58:11.826983816 +0000 UTC m=+879.911521700" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.839832 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.954904211 podStartE2EDuration="20.839813958s" podCreationTimestamp="2025-10-13 12:57:51 +0000 UTC" firstStartedPulling="2025-10-13 12:58:03.948390961 +0000 UTC m=+872.032928845" lastFinishedPulling="2025-10-13 12:58:09.833300658 +0000 UTC m=+877.917838592" observedRunningTime="2025-10-13 12:58:11.836666904 +0000 UTC m=+879.921204788" watchObservedRunningTime="2025-10-13 12:58:11.839813958 +0000 UTC m=+879.924351832" Oct 13 12:58:11 crc kubenswrapper[4678]: I1013 12:58:11.984326 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.032429 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.032664 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="dnsmasq-dns" containerID="cri-o://0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567" gracePeriod=10 Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.455173 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.555965 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc\") pod \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.556002 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config\") pod \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.556030 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm455\" (UniqueName: \"kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455\") pod \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\" (UID: \"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c\") " Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.561096 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455" (OuterVolumeSpecName: "kube-api-access-rm455") pod "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" (UID: "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c"). InnerVolumeSpecName "kube-api-access-rm455". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.598035 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" (UID: "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.603166 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config" (OuterVolumeSpecName: "config") pod "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" (UID: "61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.658524 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.658724 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.658734 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm455\" (UniqueName: \"kubernetes.io/projected/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c-kube-api-access-rm455\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.741616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4z6wg" event={"ID":"0daece72-e6c8-427f-a970-496545b49160","Type":"ContainerStarted","Data":"10168a89a9d5902600c6ac7483d0ca2c37b255de6a242e7eb60eef9003c01f2d"} Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.741683 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4z6wg" event={"ID":"0daece72-e6c8-427f-a970-496545b49160","Type":"ContainerStarted","Data":"4f940bb8192f7069b1fbc616798d3bfac136d95fbe75aedd513a74b3a6ef4394"} Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.741888 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.746557 4678 generic.go:334] "Generic (PLEG): container finished" podID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerID="0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567" exitCode=0 Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.747339 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.747791 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" event={"ID":"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c","Type":"ContainerDied","Data":"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567"} Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.747811 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-dx9js" event={"ID":"61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c","Type":"ContainerDied","Data":"44b6232f0209340c34f740b6015d48f29b4786e67b91a793715fa7ce6b18e6fd"} Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.747826 4678 scope.go:117] "RemoveContainer" containerID="0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.778343 4678 scope.go:117] "RemoveContainer" containerID="ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.778334 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-4z6wg" podStartSLOduration=11.684858795 podStartE2EDuration="16.778307764s" podCreationTimestamp="2025-10-13 12:57:56 +0000 UTC" firstStartedPulling="2025-10-13 12:58:05.239634385 +0000 UTC m=+873.324172269" lastFinishedPulling="2025-10-13 12:58:10.333083344 +0000 UTC m=+878.417621238" observedRunningTime="2025-10-13 12:58:12.761726432 +0000 UTC m=+880.846264336" watchObservedRunningTime="2025-10-13 12:58:12.778307764 +0000 UTC m=+880.862845668" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.786952 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.793588 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-dx9js"] Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.794688 4678 scope.go:117] "RemoveContainer" containerID="0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567" Oct 13 12:58:12 crc kubenswrapper[4678]: E1013 12:58:12.794951 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567\": container with ID starting with 0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567 not found: ID does not exist" containerID="0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.794978 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567"} err="failed to get container status \"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567\": rpc error: code = NotFound desc = could not find container \"0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567\": container with ID starting with 0796a263405e4bd2af7bb5945441629b58f3844cc7dacb1a6722b47abf5a9567 not found: ID does not exist" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.794997 4678 scope.go:117] "RemoveContainer" containerID="ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af" Oct 13 12:58:12 crc kubenswrapper[4678]: E1013 12:58:12.795415 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af\": container with ID starting with ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af not found: ID does not exist" containerID="ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af" Oct 13 12:58:12 crc kubenswrapper[4678]: I1013 12:58:12.795446 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af"} err="failed to get container status \"ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af\": rpc error: code = NotFound desc = could not find container \"ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af\": container with ID starting with ee7688cbdf97afa9cd50e469052a4c3500b972e6cd3906b707e37ae3d0ded8af not found: ID does not exist" Oct 13 12:58:13 crc kubenswrapper[4678]: I1013 12:58:13.756442 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:58:14 crc kubenswrapper[4678]: I1013 12:58:14.607041 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" path="/var/lib/kubelet/pods/61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c/volumes" Oct 13 12:58:14 crc kubenswrapper[4678]: I1013 12:58:14.769799 4678 generic.go:334] "Generic (PLEG): container finished" podID="7950a23a-e8d5-4d7e-a4d5-a30f535624b3" containerID="6565a13408225bff0c227367c8aaa374637ffcf6f06dce3803a9c3030714717a" exitCode=0 Oct 13 12:58:14 crc kubenswrapper[4678]: I1013 12:58:14.769956 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7950a23a-e8d5-4d7e-a4d5-a30f535624b3","Type":"ContainerDied","Data":"6565a13408225bff0c227367c8aaa374637ffcf6f06dce3803a9c3030714717a"} Oct 13 12:58:14 crc kubenswrapper[4678]: I1013 12:58:14.772498 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"16bc2692-fbba-4e51-92b4-805141199ee2","Type":"ContainerDied","Data":"a42e233121f90f03b6e714c49c682def1bea1cbcaa396d6f70cfef98371bc8a5"} Oct 13 12:58:14 crc kubenswrapper[4678]: I1013 12:58:14.772538 4678 generic.go:334] "Generic (PLEG): container finished" podID="16bc2692-fbba-4e51-92b4-805141199ee2" containerID="a42e233121f90f03b6e714c49c682def1bea1cbcaa396d6f70cfef98371bc8a5" exitCode=0 Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.786958 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7950a23a-e8d5-4d7e-a4d5-a30f535624b3","Type":"ContainerStarted","Data":"ca89f3d87d5faea97a0e1e572e0acafc6da0a0650f2f6766c135564b2406b969"} Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.790042 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"16bc2692-fbba-4e51-92b4-805141199ee2","Type":"ContainerStarted","Data":"9bbf071671a76fd823164e21633ca4dd5f88bee9cbdb325cbe323db553fc4fb5"} Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.793369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"b1da318b-1924-4ce5-9a9a-a1b483c0ffb0","Type":"ContainerStarted","Data":"5caf3a2740f8407adb4af82fe33a808100a4ae11989cdb400326b4fbb799a985"} Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.796546 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"352ce882-938a-4fe3-b4ce-7e01a5a750e6","Type":"ContainerStarted","Data":"093357e10979b8815bc7f1b7285013f72dcba2130c9ae46ec00ecdd1ef8c3ce1"} Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.826749 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.824127863 podStartE2EDuration="26.826712435s" podCreationTimestamp="2025-10-13 12:57:49 +0000 UTC" firstStartedPulling="2025-10-13 12:58:03.952308095 +0000 UTC m=+872.036845979" lastFinishedPulling="2025-10-13 12:58:09.954892667 +0000 UTC m=+878.039430551" observedRunningTime="2025-10-13 12:58:15.818007253 +0000 UTC m=+883.902545167" watchObservedRunningTime="2025-10-13 12:58:15.826712435 +0000 UTC m=+883.911250359" Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.854639 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.760830452 podStartE2EDuration="19.854610058s" podCreationTimestamp="2025-10-13 12:57:56 +0000 UTC" firstStartedPulling="2025-10-13 12:58:04.503769999 +0000 UTC m=+872.588307883" lastFinishedPulling="2025-10-13 12:58:14.597549605 +0000 UTC m=+882.682087489" observedRunningTime="2025-10-13 12:58:15.852336707 +0000 UTC m=+883.936874621" watchObservedRunningTime="2025-10-13 12:58:15.854610058 +0000 UTC m=+883.939147982" Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.889275 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.453148319 podStartE2EDuration="27.889248111s" podCreationTimestamp="2025-10-13 12:57:48 +0000 UTC" firstStartedPulling="2025-10-13 12:58:03.743870632 +0000 UTC m=+871.828408516" lastFinishedPulling="2025-10-13 12:58:10.179970434 +0000 UTC m=+878.264508308" observedRunningTime="2025-10-13 12:58:15.886858407 +0000 UTC m=+883.971396341" watchObservedRunningTime="2025-10-13 12:58:15.889248111 +0000 UTC m=+883.973786025" Oct 13 12:58:15 crc kubenswrapper[4678]: I1013 12:58:15.932927 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.831852738 podStartE2EDuration="16.932902104s" podCreationTimestamp="2025-10-13 12:57:59 +0000 UTC" firstStartedPulling="2025-10-13 12:58:05.514734135 +0000 UTC m=+873.599272019" lastFinishedPulling="2025-10-13 12:58:14.615783451 +0000 UTC m=+882.700321385" observedRunningTime="2025-10-13 12:58:15.915921371 +0000 UTC m=+884.000459315" watchObservedRunningTime="2025-10-13 12:58:15.932902104 +0000 UTC m=+884.017440018" Oct 13 12:58:16 crc kubenswrapper[4678]: I1013 12:58:16.043995 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:16 crc kubenswrapper[4678]: I1013 12:58:16.044118 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:16 crc kubenswrapper[4678]: I1013 12:58:16.115091 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:16 crc kubenswrapper[4678]: I1013 12:58:16.344384 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 13 12:58:16 crc kubenswrapper[4678]: I1013 12:58:16.854028 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.168541 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:17 crc kubenswrapper[4678]: E1013 12:58:17.175009 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="init" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.175091 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="init" Oct 13 12:58:17 crc kubenswrapper[4678]: E1013 12:58:17.175159 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="dnsmasq-dns" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.175168 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="dnsmasq-dns" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.175573 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d7d1ad-22ab-4435-be2d-7b4f13cc9c2c" containerName="dnsmasq-dns" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.176953 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.188870 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.191130 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.228768 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-67jfd"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.230038 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.240376 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.246581 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdd7j\" (UniqueName: \"kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.246659 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.246688 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.246760 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.258665 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-67jfd"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349817 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349842 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2f127fd-3a0a-44c7-9849-de6e42149fd7-config\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349867 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349918 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovs-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349962 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glfbr\" (UniqueName: \"kubernetes.io/projected/e2f127fd-3a0a-44c7-9849-de6e42149fd7-kube-api-access-glfbr\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.349979 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovn-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.350001 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-combined-ca-bundle\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.350019 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdd7j\" (UniqueName: \"kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.351068 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.351589 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.352874 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.370592 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdd7j\" (UniqueName: \"kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j\") pod \"dnsmasq-dns-7f896c8c65-kqvjv\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451505 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovs-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451562 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glfbr\" (UniqueName: \"kubernetes.io/projected/e2f127fd-3a0a-44c7-9849-de6e42149fd7-kube-api-access-glfbr\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451585 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovn-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451608 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-combined-ca-bundle\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451670 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2f127fd-3a0a-44c7-9849-de6e42149fd7-config\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.451694 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.452084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovn-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.452210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e2f127fd-3a0a-44c7-9849-de6e42149fd7-ovs-rundir\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.452965 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2f127fd-3a0a-44c7-9849-de6e42149fd7-config\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.454852 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-combined-ca-bundle\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.459478 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2f127fd-3a0a-44c7-9849-de6e42149fd7-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.469300 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glfbr\" (UniqueName: \"kubernetes.io/projected/e2f127fd-3a0a-44c7-9849-de6e42149fd7-kube-api-access-glfbr\") pod \"ovn-controller-metrics-67jfd\" (UID: \"e2f127fd-3a0a-44c7-9849-de6e42149fd7\") " pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.497284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.498629 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.501044 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.526034 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.539605 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.541971 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.557502 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.560386 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-67jfd" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.656214 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.656584 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.656617 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fppzg\" (UniqueName: \"kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.656637 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.656680 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.757998 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.758069 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fppzg\" (UniqueName: \"kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.758103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.758160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.758255 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.759314 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.759773 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.762307 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.762470 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.780603 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fppzg\" (UniqueName: \"kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg\") pod \"dnsmasq-dns-86db49b7ff-xlz6s\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.881512 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:17 crc kubenswrapper[4678]: I1013 12:58:17.956954 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:17 crc kubenswrapper[4678]: W1013 12:58:17.965024 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod718094d6_fe05_456f_b40a_c72253dd828b.slice/crio-cbc15a7927d6e7874b31b28ec2e51dc2b4318f9ccf1aa9b00e7be91c8310c0f2 WatchSource:0}: Error finding container cbc15a7927d6e7874b31b28ec2e51dc2b4318f9ccf1aa9b00e7be91c8310c0f2: Status 404 returned error can't find the container with id cbc15a7927d6e7874b31b28ec2e51dc2b4318f9ccf1aa9b00e7be91c8310c0f2 Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.074111 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-67jfd"] Oct 13 12:58:18 crc kubenswrapper[4678]: W1013 12:58:18.080886 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2f127fd_3a0a_44c7_9849_de6e42149fd7.slice/crio-bcda6f0f05a6885b1b7c8154b9aa35a89c598957a0ed9257796c9b9d6c754c17 WatchSource:0}: Error finding container bcda6f0f05a6885b1b7c8154b9aa35a89c598957a0ed9257796c9b9d6c754c17: Status 404 returned error can't find the container with id bcda6f0f05a6885b1b7c8154b9aa35a89c598957a0ed9257796c9b9d6c754c17 Oct 13 12:58:18 crc kubenswrapper[4678]: W1013 12:58:18.414726 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e8776e_1768_4cc2_977a_4671a85aa421.slice/crio-8477181df056c30de4bea68b2e686a598f17cb14ebab833eae0b361f0be60657 WatchSource:0}: Error finding container 8477181df056c30de4bea68b2e686a598f17cb14ebab833eae0b361f0be60657: Status 404 returned error can't find the container with id 8477181df056c30de4bea68b2e686a598f17cb14ebab833eae0b361f0be60657 Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.421207 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.497581 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.570005 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 13 12:58:18 crc kubenswrapper[4678]: E1013 12:58:18.704145 4678 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.30:41206->38.102.83.30:36753: write tcp 192.168.126.11:10250->192.168.126.11:37168: write: broken pipe Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.822355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67jfd" event={"ID":"e2f127fd-3a0a-44c7-9849-de6e42149fd7","Type":"ContainerStarted","Data":"5433c09d4344bd0296ea980bbd24ca3ad863e10b7631dd85d97895be76e9a881"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.822424 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-67jfd" event={"ID":"e2f127fd-3a0a-44c7-9849-de6e42149fd7","Type":"ContainerStarted","Data":"bcda6f0f05a6885b1b7c8154b9aa35a89c598957a0ed9257796c9b9d6c754c17"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.825718 4678 generic.go:334] "Generic (PLEG): container finished" podID="81e8776e-1768-4cc2-977a-4671a85aa421" containerID="071d81da841e16be34f0ef9da69c6763886429ad862ad81437f2489068a6ae97" exitCode=0 Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.825781 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" event={"ID":"81e8776e-1768-4cc2-977a-4671a85aa421","Type":"ContainerDied","Data":"071d81da841e16be34f0ef9da69c6763886429ad862ad81437f2489068a6ae97"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.825803 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" event={"ID":"81e8776e-1768-4cc2-977a-4671a85aa421","Type":"ContainerStarted","Data":"8477181df056c30de4bea68b2e686a598f17cb14ebab833eae0b361f0be60657"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.828279 4678 generic.go:334] "Generic (PLEG): container finished" podID="718094d6-fe05-456f-b40a-c72253dd828b" containerID="e8b19cd283258c7a9fbec03c057b7259aba1a3d8a5496355b947f5b420bcfed3" exitCode=0 Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.828505 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" event={"ID":"718094d6-fe05-456f-b40a-c72253dd828b","Type":"ContainerDied","Data":"e8b19cd283258c7a9fbec03c057b7259aba1a3d8a5496355b947f5b420bcfed3"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.828602 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" event={"ID":"718094d6-fe05-456f-b40a-c72253dd828b","Type":"ContainerStarted","Data":"cbc15a7927d6e7874b31b28ec2e51dc2b4318f9ccf1aa9b00e7be91c8310c0f2"} Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.847136 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-67jfd" podStartSLOduration=1.847110449 podStartE2EDuration="1.847110449s" podCreationTimestamp="2025-10-13 12:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:18.843177484 +0000 UTC m=+886.927715448" watchObservedRunningTime="2025-10-13 12:58:18.847110449 +0000 UTC m=+886.931648323" Oct 13 12:58:18 crc kubenswrapper[4678]: I1013 12:58:18.916565 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.074968 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.093705 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.093831 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.096009 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.097817 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.098225 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-dqmjg" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.102710 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.143290 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186400 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186467 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-scripts\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186504 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9kf2\" (UniqueName: \"kubernetes.io/projected/030132a0-ddd0-4001-9216-9a2962cb9f1c-kube-api-access-c9kf2\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186533 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186549 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186606 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.186631 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-config\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287370 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdd7j\" (UniqueName: \"kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j\") pod \"718094d6-fe05-456f-b40a-c72253dd828b\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287473 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc\") pod \"718094d6-fe05-456f-b40a-c72253dd828b\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287492 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config\") pod \"718094d6-fe05-456f-b40a-c72253dd828b\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287552 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb\") pod \"718094d6-fe05-456f-b40a-c72253dd828b\" (UID: \"718094d6-fe05-456f-b40a-c72253dd828b\") " Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287809 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9kf2\" (UniqueName: \"kubernetes.io/projected/030132a0-ddd0-4001-9216-9a2962cb9f1c-kube-api-access-c9kf2\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287862 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287909 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287932 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-config\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287962 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.287995 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-scripts\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.288796 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-scripts\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.289801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/030132a0-ddd0-4001-9216-9a2962cb9f1c-config\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.290068 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.293474 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j" (OuterVolumeSpecName: "kube-api-access-cdd7j") pod "718094d6-fe05-456f-b40a-c72253dd828b" (UID: "718094d6-fe05-456f-b40a-c72253dd828b"). InnerVolumeSpecName "kube-api-access-cdd7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.294208 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.294912 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.295781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/030132a0-ddd0-4001-9216-9a2962cb9f1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.314838 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9kf2\" (UniqueName: \"kubernetes.io/projected/030132a0-ddd0-4001-9216-9a2962cb9f1c-kube-api-access-c9kf2\") pod \"ovn-northd-0\" (UID: \"030132a0-ddd0-4001-9216-9a2962cb9f1c\") " pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.316212 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config" (OuterVolumeSpecName: "config") pod "718094d6-fe05-456f-b40a-c72253dd828b" (UID: "718094d6-fe05-456f-b40a-c72253dd828b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.317828 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "718094d6-fe05-456f-b40a-c72253dd828b" (UID: "718094d6-fe05-456f-b40a-c72253dd828b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.321207 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "718094d6-fe05-456f-b40a-c72253dd828b" (UID: "718094d6-fe05-456f-b40a-c72253dd828b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.389799 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.389831 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.389841 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/718094d6-fe05-456f-b40a-c72253dd828b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.389852 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdd7j\" (UniqueName: \"kubernetes.io/projected/718094d6-fe05-456f-b40a-c72253dd828b-kube-api-access-cdd7j\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.438415 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.510764 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.511213 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.588985 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.837465 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" event={"ID":"81e8776e-1768-4cc2-977a-4671a85aa421","Type":"ContainerStarted","Data":"9e91d836f2983b0d66450e0f9e46a09d965f645173d0bf1bbfd97306d5afc201"} Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.837610 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.840716 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.842241 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-kqvjv" event={"ID":"718094d6-fe05-456f-b40a-c72253dd828b","Type":"ContainerDied","Data":"cbc15a7927d6e7874b31b28ec2e51dc2b4318f9ccf1aa9b00e7be91c8310c0f2"} Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.842300 4678 scope.go:117] "RemoveContainer" containerID="e8b19cd283258c7a9fbec03c057b7259aba1a3d8a5496355b947f5b420bcfed3" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.869379 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" podStartSLOduration=2.869349086 podStartE2EDuration="2.869349086s" podCreationTimestamp="2025-10-13 12:58:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:19.85750965 +0000 UTC m=+887.942047534" watchObservedRunningTime="2025-10-13 12:58:19.869349086 +0000 UTC m=+887.953887010" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.921183 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.931338 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.978607 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:19 crc kubenswrapper[4678]: I1013 12:58:19.985962 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-kqvjv"] Oct 13 12:58:20 crc kubenswrapper[4678]: I1013 12:58:20.610667 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="718094d6-fe05-456f-b40a-c72253dd828b" path="/var/lib/kubelet/pods/718094d6-fe05-456f-b40a-c72253dd828b/volumes" Oct 13 12:58:20 crc kubenswrapper[4678]: I1013 12:58:20.852235 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"030132a0-ddd0-4001-9216-9a2962cb9f1c","Type":"ContainerStarted","Data":"6565a56e2887ef521dd0feccd5e4adf1629878b43dd101149de603eb82b32e0e"} Oct 13 12:58:20 crc kubenswrapper[4678]: I1013 12:58:20.903276 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 13 12:58:20 crc kubenswrapper[4678]: I1013 12:58:20.903580 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.055386 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vgs9f"] Oct 13 12:58:21 crc kubenswrapper[4678]: E1013 12:58:21.055767 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="718094d6-fe05-456f-b40a-c72253dd828b" containerName="init" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.055786 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="718094d6-fe05-456f-b40a-c72253dd828b" containerName="init" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.056107 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="718094d6-fe05-456f-b40a-c72253dd828b" containerName="init" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.057670 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.063596 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vgs9f"] Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.122593 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72297\" (UniqueName: \"kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297\") pod \"keystone-db-create-vgs9f\" (UID: \"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159\") " pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.223995 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72297\" (UniqueName: \"kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297\") pod \"keystone-db-create-vgs9f\" (UID: \"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159\") " pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.249450 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72297\" (UniqueName: \"kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297\") pod \"keystone-db-create-vgs9f\" (UID: \"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159\") " pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.419946 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-d4hbz"] Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.422014 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.426407 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4hbz"] Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.473759 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.529491 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzd2r\" (UniqueName: \"kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r\") pod \"placement-db-create-d4hbz\" (UID: \"9ac932bc-64cc-4681-b720-dc6f2ee45f6c\") " pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.631600 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzd2r\" (UniqueName: \"kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r\") pod \"placement-db-create-d4hbz\" (UID: \"9ac932bc-64cc-4681-b720-dc6f2ee45f6c\") " pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.650212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzd2r\" (UniqueName: \"kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r\") pod \"placement-db-create-d4hbz\" (UID: \"9ac932bc-64cc-4681-b720-dc6f2ee45f6c\") " pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.741128 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.860698 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"030132a0-ddd0-4001-9216-9a2962cb9f1c","Type":"ContainerStarted","Data":"0c66bdce9402ee4f45ad4f878ee99c14588277c721a195e6f897f9da59a62962"} Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.862188 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"030132a0-ddd0-4001-9216-9a2962cb9f1c","Type":"ContainerStarted","Data":"83655c786ebf50e88da86e8d7ffd06d240bc321dd727aa4df97cb18028206b5f"} Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.862274 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.891343 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.818915415 podStartE2EDuration="2.891283387s" podCreationTimestamp="2025-10-13 12:58:19 +0000 UTC" firstStartedPulling="2025-10-13 12:58:19.942042043 +0000 UTC m=+888.026579927" lastFinishedPulling="2025-10-13 12:58:21.014410015 +0000 UTC m=+889.098947899" observedRunningTime="2025-10-13 12:58:21.882397441 +0000 UTC m=+889.966935325" watchObservedRunningTime="2025-10-13 12:58:21.891283387 +0000 UTC m=+889.975821291" Oct 13 12:58:21 crc kubenswrapper[4678]: I1013 12:58:21.969737 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vgs9f"] Oct 13 12:58:21 crc kubenswrapper[4678]: W1013 12:58:21.971795 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fae8bdb_cfa7_4d91_bd3c_ddd8efc2d159.slice/crio-448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f WatchSource:0}: Error finding container 448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f: Status 404 returned error can't find the container with id 448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f Oct 13 12:58:22 crc kubenswrapper[4678]: I1013 12:58:22.183401 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4hbz"] Oct 13 12:58:22 crc kubenswrapper[4678]: W1013 12:58:22.185295 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ac932bc_64cc_4681_b720_dc6f2ee45f6c.slice/crio-76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4 WatchSource:0}: Error finding container 76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4: Status 404 returned error can't find the container with id 76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4 Oct 13 12:58:22 crc kubenswrapper[4678]: I1013 12:58:22.873708 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vgs9f" event={"ID":"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159","Type":"ContainerStarted","Data":"448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f"} Oct 13 12:58:22 crc kubenswrapper[4678]: I1013 12:58:22.876682 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4hbz" event={"ID":"9ac932bc-64cc-4681-b720-dc6f2ee45f6c","Type":"ContainerStarted","Data":"76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4"} Oct 13 12:58:22 crc kubenswrapper[4678]: I1013 12:58:22.972616 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.049970 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.379664 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.458033 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.458261 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="dnsmasq-dns" containerID="cri-o://9e91d836f2983b0d66450e0f9e46a09d965f645173d0bf1bbfd97306d5afc201" gracePeriod=10 Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.481717 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.482841 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.499314 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.586768 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.586811 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.586843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.586901 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.586932 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr7h6\" (UniqueName: \"kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.688386 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr7h6\" (UniqueName: \"kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.688765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.688788 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.688829 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.688884 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.689647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.689663 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.730289 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.730648 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.738095 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr7h6\" (UniqueName: \"kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6\") pod \"dnsmasq-dns-698758b865-2clp7\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.796479 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.901912 4678 generic.go:334] "Generic (PLEG): container finished" podID="81e8776e-1768-4cc2-977a-4671a85aa421" containerID="9e91d836f2983b0d66450e0f9e46a09d965f645173d0bf1bbfd97306d5afc201" exitCode=0 Oct 13 12:58:23 crc kubenswrapper[4678]: I1013 12:58:23.902118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" event={"ID":"81e8776e-1768-4cc2-977a-4671a85aa421","Type":"ContainerDied","Data":"9e91d836f2983b0d66450e0f9e46a09d965f645173d0bf1bbfd97306d5afc201"} Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.274325 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.616619 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.634156 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.634337 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.636224 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.638295 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.638340 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-7gl52" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.638426 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.805870 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-lock\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.805931 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.805992 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvbxv\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-kube-api-access-xvbxv\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.806083 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.806162 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-cache\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908240 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908365 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-cache\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908495 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-lock\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908543 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908607 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvbxv\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-kube-api-access-xvbxv\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: E1013 12:58:24.908742 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 12:58:24 crc kubenswrapper[4678]: E1013 12:58:24.908780 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 12:58:24 crc kubenswrapper[4678]: E1013 12:58:24.908854 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift podName:e2337ebf-ba94-424a-9b45-fb68453bbf64 nodeName:}" failed. No retries permitted until 2025-10-13 12:58:25.408831517 +0000 UTC m=+893.493369441 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift") pod "swift-storage-0" (UID: "e2337ebf-ba94-424a-9b45-fb68453bbf64") : configmap "swift-ring-files" not found Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.908914 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-cache\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.909128 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e2337ebf-ba94-424a-9b45-fb68453bbf64-lock\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.909151 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.911099 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2clp7" event={"ID":"f5258b79-3ee1-4989-bc3c-23711b14f1a6","Type":"ContainerStarted","Data":"c2453a2760ceb1998afb914eb5c431fa647e06b46f80c1f4dd93b53c5996e0a4"} Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.932193 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvbxv\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-kube-api-access-xvbxv\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:24 crc kubenswrapper[4678]: I1013 12:58:24.946372 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:25 crc kubenswrapper[4678]: I1013 12:58:25.416350 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:25 crc kubenswrapper[4678]: E1013 12:58:25.416530 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 12:58:25 crc kubenswrapper[4678]: E1013 12:58:25.416552 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 12:58:25 crc kubenswrapper[4678]: E1013 12:58:25.416600 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift podName:e2337ebf-ba94-424a-9b45-fb68453bbf64 nodeName:}" failed. No retries permitted until 2025-10-13 12:58:26.416586415 +0000 UTC m=+894.501124299 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift") pod "swift-storage-0" (UID: "e2337ebf-ba94-424a-9b45-fb68453bbf64") : configmap "swift-ring-files" not found Oct 13 12:58:25 crc kubenswrapper[4678]: I1013 12:58:25.506457 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:58:25 crc kubenswrapper[4678]: I1013 12:58:25.506537 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.433266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:26 crc kubenswrapper[4678]: E1013 12:58:26.433569 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 12:58:26 crc kubenswrapper[4678]: E1013 12:58:26.433814 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 12:58:26 crc kubenswrapper[4678]: E1013 12:58:26.433918 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift podName:e2337ebf-ba94-424a-9b45-fb68453bbf64 nodeName:}" failed. No retries permitted until 2025-10-13 12:58:28.43388493 +0000 UTC m=+896.518422854 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift") pod "swift-storage-0" (UID: "e2337ebf-ba94-424a-9b45-fb68453bbf64") : configmap "swift-ring-files" not found Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.629809 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-dv4hj"] Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.630952 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.646742 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dv4hj"] Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.739293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x82g\" (UniqueName: \"kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g\") pod \"glance-db-create-dv4hj\" (UID: \"80039354-df5b-44f1-a7c5-1b997e49229a\") " pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.840990 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x82g\" (UniqueName: \"kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g\") pod \"glance-db-create-dv4hj\" (UID: \"80039354-df5b-44f1-a7c5-1b997e49229a\") " pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.869332 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x82g\" (UniqueName: \"kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g\") pod \"glance-db-create-dv4hj\" (UID: \"80039354-df5b-44f1-a7c5-1b997e49229a\") " pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.934688 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vgs9f" event={"ID":"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159","Type":"ContainerStarted","Data":"4224014be7090eb5eb60ade9164db368e982b89cc74f8221f1dc639f8ee9fc17"} Oct 13 12:58:26 crc kubenswrapper[4678]: I1013 12:58:26.957717 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.436240 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-dv4hj"] Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.616142 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.760648 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb\") pod \"81e8776e-1768-4cc2-977a-4671a85aa421\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.760686 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb\") pod \"81e8776e-1768-4cc2-977a-4671a85aa421\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.760733 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc\") pod \"81e8776e-1768-4cc2-977a-4671a85aa421\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.760787 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fppzg\" (UniqueName: \"kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg\") pod \"81e8776e-1768-4cc2-977a-4671a85aa421\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.760839 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config\") pod \"81e8776e-1768-4cc2-977a-4671a85aa421\" (UID: \"81e8776e-1768-4cc2-977a-4671a85aa421\") " Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.765844 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg" (OuterVolumeSpecName: "kube-api-access-fppzg") pod "81e8776e-1768-4cc2-977a-4671a85aa421" (UID: "81e8776e-1768-4cc2-977a-4671a85aa421"). InnerVolumeSpecName "kube-api-access-fppzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.795868 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "81e8776e-1768-4cc2-977a-4671a85aa421" (UID: "81e8776e-1768-4cc2-977a-4671a85aa421"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.806986 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "81e8776e-1768-4cc2-977a-4671a85aa421" (UID: "81e8776e-1768-4cc2-977a-4671a85aa421"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.811440 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config" (OuterVolumeSpecName: "config") pod "81e8776e-1768-4cc2-977a-4671a85aa421" (UID: "81e8776e-1768-4cc2-977a-4671a85aa421"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.814582 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "81e8776e-1768-4cc2-977a-4671a85aa421" (UID: "81e8776e-1768-4cc2-977a-4671a85aa421"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.863005 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.863051 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.863089 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.863104 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/81e8776e-1768-4cc2-977a-4671a85aa421-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.863116 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fppzg\" (UniqueName: \"kubernetes.io/projected/81e8776e-1768-4cc2-977a-4671a85aa421-kube-api-access-fppzg\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.943555 4678 generic.go:334] "Generic (PLEG): container finished" podID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerID="3c6511b004faf36fb28e55e449c0f68595773a33da06e995c4976439434c5085" exitCode=0 Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.943616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2clp7" event={"ID":"f5258b79-3ee1-4989-bc3c-23711b14f1a6","Type":"ContainerDied","Data":"3c6511b004faf36fb28e55e449c0f68595773a33da06e995c4976439434c5085"} Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.945723 4678 generic.go:334] "Generic (PLEG): container finished" podID="9ac932bc-64cc-4681-b720-dc6f2ee45f6c" containerID="dc471b38e88b76899004ebd4681428dfcb9b2ddb75edc5c151eb2c49ddd7ed06" exitCode=0 Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.945830 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4hbz" event={"ID":"9ac932bc-64cc-4681-b720-dc6f2ee45f6c","Type":"ContainerDied","Data":"dc471b38e88b76899004ebd4681428dfcb9b2ddb75edc5c151eb2c49ddd7ed06"} Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.947569 4678 generic.go:334] "Generic (PLEG): container finished" podID="3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" containerID="4224014be7090eb5eb60ade9164db368e982b89cc74f8221f1dc639f8ee9fc17" exitCode=0 Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.947625 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vgs9f" event={"ID":"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159","Type":"ContainerDied","Data":"4224014be7090eb5eb60ade9164db368e982b89cc74f8221f1dc639f8ee9fc17"} Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.950728 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" event={"ID":"81e8776e-1768-4cc2-977a-4671a85aa421","Type":"ContainerDied","Data":"8477181df056c30de4bea68b2e686a598f17cb14ebab833eae0b361f0be60657"} Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.950767 4678 scope.go:117] "RemoveContainer" containerID="9e91d836f2983b0d66450e0f9e46a09d965f645173d0bf1bbfd97306d5afc201" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.950872 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xlz6s" Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.957960 4678 generic.go:334] "Generic (PLEG): container finished" podID="80039354-df5b-44f1-a7c5-1b997e49229a" containerID="56da00d5f38771b2ff822dc867eeb009b4feb7906cd960cfd1516c7ca70c7298" exitCode=0 Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.958000 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dv4hj" event={"ID":"80039354-df5b-44f1-a7c5-1b997e49229a","Type":"ContainerDied","Data":"56da00d5f38771b2ff822dc867eeb009b4feb7906cd960cfd1516c7ca70c7298"} Oct 13 12:58:27 crc kubenswrapper[4678]: I1013 12:58:27.958027 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dv4hj" event={"ID":"80039354-df5b-44f1-a7c5-1b997e49229a","Type":"ContainerStarted","Data":"8ddc1fce2936120e5e241a3c010339bce35b69247545950e8fe8494f3b886e7f"} Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.069258 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.077394 4678 scope.go:117] "RemoveContainer" containerID="071d81da841e16be34f0ef9da69c6763886429ad862ad81437f2489068a6ae97" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.079485 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xlz6s"] Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.482641 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.483057 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.483123 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.483210 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift podName:e2337ebf-ba94-424a-9b45-fb68453bbf64 nodeName:}" failed. No retries permitted until 2025-10-13 12:58:32.483180281 +0000 UTC m=+900.567718215 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift") pod "swift-storage-0" (UID: "e2337ebf-ba94-424a-9b45-fb68453bbf64") : configmap "swift-ring-files" not found Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.557480 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6s9lw"] Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.558291 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="init" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.558428 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="init" Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.558555 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="dnsmasq-dns" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.558662 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="dnsmasq-dns" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.559084 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" containerName="dnsmasq-dns" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.560160 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.564098 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.564554 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.571705 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.604427 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e8776e-1768-4cc2-977a-4671a85aa421" path="/var/lib/kubelet/pods/81e8776e-1768-4cc2-977a-4671a85aa421/volumes" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.606537 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6s9lw"] Oct 13 12:58:28 crc kubenswrapper[4678]: E1013 12:58:28.607199 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-67j98 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-6s9lw" podUID="000f47ee-4102-4c68-a938-8bd1cc18eeab" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.614900 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6x2jw"] Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.616242 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.626687 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6s9lw"] Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.645617 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6x2jw"] Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.686464 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.686564 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.686642 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.686768 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67j98\" (UniqueName: \"kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.686816 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.687219 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.687402 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788705 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788770 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67j98\" (UniqueName: \"kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788837 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788870 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788917 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.788970 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789016 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789093 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-std8d\" (UniqueName: \"kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789215 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789255 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789310 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789343 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.789377 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.790987 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.791299 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.791546 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.794624 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.794748 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.795046 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.816337 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67j98\" (UniqueName: \"kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98\") pod \"swift-ring-rebalance-6s9lw\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.891920 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.892973 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.893273 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.893521 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.893624 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.894249 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.894468 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.894655 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.894792 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-std8d\" (UniqueName: \"kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.896164 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.897539 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.900197 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.902668 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.939196 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-std8d\" (UniqueName: \"kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d\") pod \"swift-ring-rebalance-6x2jw\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.970370 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2clp7" event={"ID":"f5258b79-3ee1-4989-bc3c-23711b14f1a6","Type":"ContainerStarted","Data":"85f9bfa30f288b130591f59db2a145fa3af223ee4eed2591ca3d0c91d2f2bad2"} Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.971708 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:28 crc kubenswrapper[4678]: I1013 12:58:28.997909 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.005800 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-2clp7" podStartSLOduration=6.005769884 podStartE2EDuration="6.005769884s" podCreationTimestamp="2025-10-13 12:58:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:28.992350687 +0000 UTC m=+897.076888591" watchObservedRunningTime="2025-10-13 12:58:29.005769884 +0000 UTC m=+897.090307788" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098198 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098298 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098334 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098365 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098445 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098518 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67j98\" (UniqueName: \"kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.098641 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf\") pod \"000f47ee-4102-4c68-a938-8bd1cc18eeab\" (UID: \"000f47ee-4102-4c68-a938-8bd1cc18eeab\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.100104 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.100119 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts" (OuterVolumeSpecName: "scripts") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.100333 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.104241 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.105006 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.105402 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.118613 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98" (OuterVolumeSpecName: "kube-api-access-67j98") pod "000f47ee-4102-4c68-a938-8bd1cc18eeab" (UID: "000f47ee-4102-4c68-a938-8bd1cc18eeab"). InnerVolumeSpecName "kube-api-access-67j98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201344 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201384 4678 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/000f47ee-4102-4c68-a938-8bd1cc18eeab-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201399 4678 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201411 4678 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201424 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/000f47ee-4102-4c68-a938-8bd1cc18eeab-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201436 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67j98\" (UniqueName: \"kubernetes.io/projected/000f47ee-4102-4c68-a938-8bd1cc18eeab-kube-api-access-67j98\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.201449 4678 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/000f47ee-4102-4c68-a938-8bd1cc18eeab-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.239266 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.304400 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.381713 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.400987 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.404300 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzd2r\" (UniqueName: \"kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r\") pod \"9ac932bc-64cc-4681-b720-dc6f2ee45f6c\" (UID: \"9ac932bc-64cc-4681-b720-dc6f2ee45f6c\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.408904 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r" (OuterVolumeSpecName: "kube-api-access-mzd2r") pod "9ac932bc-64cc-4681-b720-dc6f2ee45f6c" (UID: "9ac932bc-64cc-4681-b720-dc6f2ee45f6c"). InnerVolumeSpecName "kube-api-access-mzd2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.506092 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72297\" (UniqueName: \"kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297\") pod \"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159\" (UID: \"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.506261 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x82g\" (UniqueName: \"kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g\") pod \"80039354-df5b-44f1-a7c5-1b997e49229a\" (UID: \"80039354-df5b-44f1-a7c5-1b997e49229a\") " Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.506983 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzd2r\" (UniqueName: \"kubernetes.io/projected/9ac932bc-64cc-4681-b720-dc6f2ee45f6c-kube-api-access-mzd2r\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.510634 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g" (OuterVolumeSpecName: "kube-api-access-7x82g") pod "80039354-df5b-44f1-a7c5-1b997e49229a" (UID: "80039354-df5b-44f1-a7c5-1b997e49229a"). InnerVolumeSpecName "kube-api-access-7x82g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.510700 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297" (OuterVolumeSpecName: "kube-api-access-72297") pod "3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" (UID: "3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159"). InnerVolumeSpecName "kube-api-access-72297". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.608640 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72297\" (UniqueName: \"kubernetes.io/projected/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159-kube-api-access-72297\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.608673 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x82g\" (UniqueName: \"kubernetes.io/projected/80039354-df5b-44f1-a7c5-1b997e49229a-kube-api-access-7x82g\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.725810 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6x2jw"] Oct 13 12:58:29 crc kubenswrapper[4678]: W1013 12:58:29.730914 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07587cb9_f0d2_41d5_a596_7650a30f9cd1.slice/crio-b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696 WatchSource:0}: Error finding container b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696: Status 404 returned error can't find the container with id b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696 Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.978955 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-dv4hj" event={"ID":"80039354-df5b-44f1-a7c5-1b997e49229a","Type":"ContainerDied","Data":"8ddc1fce2936120e5e241a3c010339bce35b69247545950e8fe8494f3b886e7f"} Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.979367 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ddc1fce2936120e5e241a3c010339bce35b69247545950e8fe8494f3b886e7f" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.979012 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-dv4hj" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.980474 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6x2jw" event={"ID":"07587cb9-f0d2-41d5-a596-7650a30f9cd1","Type":"ContainerStarted","Data":"b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696"} Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.981779 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4hbz" event={"ID":"9ac932bc-64cc-4681-b720-dc6f2ee45f6c","Type":"ContainerDied","Data":"76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4"} Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.981803 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76a2b84b0f09544d7a8983b436835c048abd1c4b168479e243be7db4fec6e1e4" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.981856 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4hbz" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.982918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vgs9f" event={"ID":"3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159","Type":"ContainerDied","Data":"448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f"} Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.982949 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="448492d44bb10b0635b9c408cc6b536869ba5b3b509c82e186aa7700acc3c54f" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.982932 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6s9lw" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.983001 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vgs9f" Oct 13 12:58:29 crc kubenswrapper[4678]: I1013 12:58:29.983081 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:30 crc kubenswrapper[4678]: I1013 12:58:30.043342 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-6s9lw"] Oct 13 12:58:30 crc kubenswrapper[4678]: I1013 12:58:30.054561 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-6s9lw"] Oct 13 12:58:30 crc kubenswrapper[4678]: I1013 12:58:30.607438 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000f47ee-4102-4c68-a938-8bd1cc18eeab" path="/var/lib/kubelet/pods/000f47ee-4102-4c68-a938-8bd1cc18eeab/volumes" Oct 13 12:58:32 crc kubenswrapper[4678]: I1013 12:58:32.560775 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:32 crc kubenswrapper[4678]: E1013 12:58:32.561529 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 13 12:58:32 crc kubenswrapper[4678]: E1013 12:58:32.561697 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 13 12:58:32 crc kubenswrapper[4678]: E1013 12:58:32.561776 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift podName:e2337ebf-ba94-424a-9b45-fb68453bbf64 nodeName:}" failed. No retries permitted until 2025-10-13 12:58:40.561751489 +0000 UTC m=+908.646289403 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift") pod "swift-storage-0" (UID: "e2337ebf-ba94-424a-9b45-fb68453bbf64") : configmap "swift-ring-files" not found Oct 13 12:58:33 crc kubenswrapper[4678]: I1013 12:58:33.024093 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6x2jw" event={"ID":"07587cb9-f0d2-41d5-a596-7650a30f9cd1","Type":"ContainerStarted","Data":"2ca61678744f2185800339f36c3c94dcdbb9baff8b50750ef6c99ebae46ade5d"} Oct 13 12:58:33 crc kubenswrapper[4678]: I1013 12:58:33.051150 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6x2jw" podStartSLOduration=2.200409454 podStartE2EDuration="5.051132368s" podCreationTimestamp="2025-10-13 12:58:28 +0000 UTC" firstStartedPulling="2025-10-13 12:58:29.733086543 +0000 UTC m=+897.817624437" lastFinishedPulling="2025-10-13 12:58:32.583809427 +0000 UTC m=+900.668347351" observedRunningTime="2025-10-13 12:58:33.043437203 +0000 UTC m=+901.127975097" watchObservedRunningTime="2025-10-13 12:58:33.051132368 +0000 UTC m=+901.135670262" Oct 13 12:58:33 crc kubenswrapper[4678]: I1013 12:58:33.798334 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:58:33 crc kubenswrapper[4678]: I1013 12:58:33.870536 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:58:33 crc kubenswrapper[4678]: I1013 12:58:33.871263 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="dnsmasq-dns" containerID="cri-o://6373bdec25a6156f7e78841351fde5860bc32a8ccfe52f76e7bdaba502a33466" gracePeriod=10 Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.036328 4678 generic.go:334] "Generic (PLEG): container finished" podID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerID="6373bdec25a6156f7e78841351fde5860bc32a8ccfe52f76e7bdaba502a33466" exitCode=0 Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.036382 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" event={"ID":"390d51cb-5775-4cc2-85c9-a2cd23cb9144","Type":"ContainerDied","Data":"6373bdec25a6156f7e78841351fde5860bc32a8ccfe52f76e7bdaba502a33466"} Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.348529 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.496261 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.504920 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config\") pod \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.504973 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc\") pod \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.505004 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch9wg\" (UniqueName: \"kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg\") pod \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\" (UID: \"390d51cb-5775-4cc2-85c9-a2cd23cb9144\") " Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.511583 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg" (OuterVolumeSpecName: "kube-api-access-ch9wg") pod "390d51cb-5775-4cc2-85c9-a2cd23cb9144" (UID: "390d51cb-5775-4cc2-85c9-a2cd23cb9144"). InnerVolumeSpecName "kube-api-access-ch9wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.556157 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config" (OuterVolumeSpecName: "config") pod "390d51cb-5775-4cc2-85c9-a2cd23cb9144" (UID: "390d51cb-5775-4cc2-85c9-a2cd23cb9144"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.562176 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "390d51cb-5775-4cc2-85c9-a2cd23cb9144" (UID: "390d51cb-5775-4cc2-85c9-a2cd23cb9144"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.606774 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.606809 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/390d51cb-5775-4cc2-85c9-a2cd23cb9144-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:34 crc kubenswrapper[4678]: I1013 12:58:34.606821 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch9wg\" (UniqueName: \"kubernetes.io/projected/390d51cb-5775-4cc2-85c9-a2cd23cb9144-kube-api-access-ch9wg\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.039930 4678 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod11588d48-2169-423c-b1c3-597795548526"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod11588d48-2169-423c-b1c3-597795548526] : Timed out while waiting for systemd to remove kubepods-besteffort-pod11588d48_2169_423c_b1c3_597795548526.slice" Oct 13 12:58:35 crc kubenswrapper[4678]: E1013 12:58:35.039970 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod11588d48-2169-423c-b1c3-597795548526] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod11588d48-2169-423c-b1c3-597795548526] : Timed out while waiting for systemd to remove kubepods-besteffort-pod11588d48_2169_423c_b1c3_597795548526.slice" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" podUID="11588d48-2169-423c-b1c3-597795548526" Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.044535 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" event={"ID":"390d51cb-5775-4cc2-85c9-a2cd23cb9144","Type":"ContainerDied","Data":"3620a7996f7785142af97805a68e509543a5ab21761ad7e7eb1406c6e0c265dd"} Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.044586 4678 scope.go:117] "RemoveContainer" containerID="6373bdec25a6156f7e78841351fde5860bc32a8ccfe52f76e7bdaba502a33466" Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.044713 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2cl4" Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.079532 4678 scope.go:117] "RemoveContainer" containerID="271f9e7dea4d7cd9e78c72888fb67d9ceb9defdd718460124052fb23efedfd9b" Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.092277 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:58:35 crc kubenswrapper[4678]: I1013 12:58:35.105589 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2cl4"] Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.061148 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xgsfr" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.162575 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.192085 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xgsfr"] Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.608982 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11588d48-2169-423c-b1c3-597795548526" path="/var/lib/kubelet/pods/11588d48-2169-423c-b1c3-597795548526/volumes" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.609808 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" path="/var/lib/kubelet/pods/390d51cb-5775-4cc2-85c9-a2cd23cb9144/volumes" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.645875 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9be8-account-create-t74kb"] Oct 13 12:58:36 crc kubenswrapper[4678]: E1013 12:58:36.646493 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="dnsmasq-dns" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646522 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="dnsmasq-dns" Oct 13 12:58:36 crc kubenswrapper[4678]: E1013 12:58:36.646543 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac932bc-64cc-4681-b720-dc6f2ee45f6c" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646556 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac932bc-64cc-4681-b720-dc6f2ee45f6c" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: E1013 12:58:36.646587 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80039354-df5b-44f1-a7c5-1b997e49229a" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646600 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="80039354-df5b-44f1-a7c5-1b997e49229a" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: E1013 12:58:36.646623 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646637 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: E1013 12:58:36.646657 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="init" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646669 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="init" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646964 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="390d51cb-5775-4cc2-85c9-a2cd23cb9144" containerName="dnsmasq-dns" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.646995 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.647011 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="80039354-df5b-44f1-a7c5-1b997e49229a" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.647038 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac932bc-64cc-4681-b720-dc6f2ee45f6c" containerName="mariadb-database-create" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.647918 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.651264 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.656427 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9be8-account-create-t74kb"] Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.743711 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggxfs\" (UniqueName: \"kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs\") pod \"glance-9be8-account-create-t74kb\" (UID: \"c6d59036-e487-4f27-9522-4db47fe067f5\") " pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.846089 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggxfs\" (UniqueName: \"kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs\") pod \"glance-9be8-account-create-t74kb\" (UID: \"c6d59036-e487-4f27-9522-4db47fe067f5\") " pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.877276 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggxfs\" (UniqueName: \"kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs\") pod \"glance-9be8-account-create-t74kb\" (UID: \"c6d59036-e487-4f27-9522-4db47fe067f5\") " pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:36 crc kubenswrapper[4678]: I1013 12:58:36.976248 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:37 crc kubenswrapper[4678]: I1013 12:58:37.119225 4678 generic.go:334] "Generic (PLEG): container finished" podID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerID="0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a" exitCode=0 Oct 13 12:58:37 crc kubenswrapper[4678]: I1013 12:58:37.119276 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerDied","Data":"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a"} Oct 13 12:58:37 crc kubenswrapper[4678]: I1013 12:58:37.509296 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9be8-account-create-t74kb"] Oct 13 12:58:37 crc kubenswrapper[4678]: W1013 12:58:37.511462 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6d59036_e487_4f27_9522_4db47fe067f5.slice/crio-0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae WatchSource:0}: Error finding container 0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae: Status 404 returned error can't find the container with id 0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.130923 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerStarted","Data":"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5"} Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.131575 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.132419 4678 generic.go:334] "Generic (PLEG): container finished" podID="c6d59036-e487-4f27-9522-4db47fe067f5" containerID="2932c5ee6c30cecce0d821a793803ace56c0cca8858436bbb85ba77df531bd83" exitCode=0 Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.132467 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9be8-account-create-t74kb" event={"ID":"c6d59036-e487-4f27-9522-4db47fe067f5","Type":"ContainerDied","Data":"2932c5ee6c30cecce0d821a793803ace56c0cca8858436bbb85ba77df531bd83"} Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.132486 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9be8-account-create-t74kb" event={"ID":"c6d59036-e487-4f27-9522-4db47fe067f5","Type":"ContainerStarted","Data":"0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae"} Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.134561 4678 generic.go:334] "Generic (PLEG): container finished" podID="d90b147e-003a-4981-ab10-3e933cf4be70" containerID="a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf" exitCode=0 Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.134605 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerDied","Data":"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf"} Oct 13 12:58:38 crc kubenswrapper[4678]: I1013 12:58:38.161587 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.50503746 podStartE2EDuration="52.161563611s" podCreationTimestamp="2025-10-13 12:57:46 +0000 UTC" firstStartedPulling="2025-10-13 12:57:52.745807812 +0000 UTC m=+860.830345686" lastFinishedPulling="2025-10-13 12:58:03.402333953 +0000 UTC m=+871.486871837" observedRunningTime="2025-10-13 12:58:38.160970166 +0000 UTC m=+906.245508050" watchObservedRunningTime="2025-10-13 12:58:38.161563611 +0000 UTC m=+906.246101505" Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.144920 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerStarted","Data":"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91"} Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.146097 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.147128 4678 generic.go:334] "Generic (PLEG): container finished" podID="07587cb9-f0d2-41d5-a596-7650a30f9cd1" containerID="2ca61678744f2185800339f36c3c94dcdbb9baff8b50750ef6c99ebae46ade5d" exitCode=0 Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.147235 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6x2jw" event={"ID":"07587cb9-f0d2-41d5-a596-7650a30f9cd1","Type":"ContainerDied","Data":"2ca61678744f2185800339f36c3c94dcdbb9baff8b50750ef6c99ebae46ade5d"} Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.166163 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.166124577 podStartE2EDuration="53.166124577s" podCreationTimestamp="2025-10-13 12:57:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:39.163104297 +0000 UTC m=+907.247642191" watchObservedRunningTime="2025-10-13 12:58:39.166124577 +0000 UTC m=+907.250662471" Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.495973 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.596247 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggxfs\" (UniqueName: \"kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs\") pod \"c6d59036-e487-4f27-9522-4db47fe067f5\" (UID: \"c6d59036-e487-4f27-9522-4db47fe067f5\") " Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.611314 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs" (OuterVolumeSpecName: "kube-api-access-ggxfs") pod "c6d59036-e487-4f27-9522-4db47fe067f5" (UID: "c6d59036-e487-4f27-9522-4db47fe067f5"). InnerVolumeSpecName "kube-api-access-ggxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:39 crc kubenswrapper[4678]: I1013 12:58:39.700194 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggxfs\" (UniqueName: \"kubernetes.io/projected/c6d59036-e487-4f27-9522-4db47fe067f5-kube-api-access-ggxfs\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.155580 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9be8-account-create-t74kb" event={"ID":"c6d59036-e487-4f27-9522-4db47fe067f5","Type":"ContainerDied","Data":"0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae"} Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.156338 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ca1986971b7fb5c5782ca513938c7d0ebd2cbc48a10051206f5c71de26edcae" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.156524 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9be8-account-create-t74kb" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.551074 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614450 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614530 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614558 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-std8d\" (UniqueName: \"kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614593 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614610 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614643 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614661 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices\") pod \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\" (UID: \"07587cb9-f0d2-41d5-a596-7650a30f9cd1\") " Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.614955 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.615601 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.615925 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.619963 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e2337ebf-ba94-424a-9b45-fb68453bbf64-etc-swift\") pod \"swift-storage-0\" (UID: \"e2337ebf-ba94-424a-9b45-fb68453bbf64\") " pod="openstack/swift-storage-0" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.637420 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.649025 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d" (OuterVolumeSpecName: "kube-api-access-std8d") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "kube-api-access-std8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.657968 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.659380 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.674542 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts" (OuterVolumeSpecName: "scripts") pod "07587cb9-f0d2-41d5-a596-7650a30f9cd1" (UID: "07587cb9-f0d2-41d5-a596-7650a30f9cd1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.716845 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717185 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-std8d\" (UniqueName: \"kubernetes.io/projected/07587cb9-f0d2-41d5-a596-7650a30f9cd1-kube-api-access-std8d\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717281 4678 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717364 4678 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/07587cb9-f0d2-41d5-a596-7650a30f9cd1-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717456 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717560 4678 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/07587cb9-f0d2-41d5-a596-7650a30f9cd1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.717642 4678 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/07587cb9-f0d2-41d5-a596-7650a30f9cd1-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:40 crc kubenswrapper[4678]: I1013 12:58:40.855887 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.080032 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5e8c-account-create-dhm4c"] Oct 13 12:58:41 crc kubenswrapper[4678]: E1013 12:58:41.080596 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6d59036-e487-4f27-9522-4db47fe067f5" containerName="mariadb-account-create" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.080620 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6d59036-e487-4f27-9522-4db47fe067f5" containerName="mariadb-account-create" Oct 13 12:58:41 crc kubenswrapper[4678]: E1013 12:58:41.080657 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07587cb9-f0d2-41d5-a596-7650a30f9cd1" containerName="swift-ring-rebalance" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.080667 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="07587cb9-f0d2-41d5-a596-7650a30f9cd1" containerName="swift-ring-rebalance" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.080946 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6d59036-e487-4f27-9522-4db47fe067f5" containerName="mariadb-account-create" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.080974 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="07587cb9-f0d2-41d5-a596-7650a30f9cd1" containerName="swift-ring-rebalance" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.082071 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.085661 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.090647 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5e8c-account-create-dhm4c"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.123519 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpc5c\" (UniqueName: \"kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c\") pod \"keystone-5e8c-account-create-dhm4c\" (UID: \"d624a6f3-db83-4d66-b86a-0d76348a9839\") " pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.166667 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6x2jw" event={"ID":"07587cb9-f0d2-41d5-a596-7650a30f9cd1","Type":"ContainerDied","Data":"b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696"} Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.166703 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b250cc981cda0495102568a4ac1f3640798d64f2ad2644df6a556de62345f696" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.166751 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6x2jw" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.225156 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpc5c\" (UniqueName: \"kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c\") pod \"keystone-5e8c-account-create-dhm4c\" (UID: \"d624a6f3-db83-4d66-b86a-0d76348a9839\") " pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.248829 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpc5c\" (UniqueName: \"kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c\") pod \"keystone-5e8c-account-create-dhm4c\" (UID: \"d624a6f3-db83-4d66-b86a-0d76348a9839\") " pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.411832 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.474230 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-96f2-account-create-4h6m4"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.475189 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.477946 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.533246 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96f2-account-create-4h6m4"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.549629 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.632763 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-dv2nq" podUID="2e724ee3-a6c7-4b75-9c60-3fe2f233b96b" containerName="ovn-controller" probeResult="failure" output=< Oct 13 12:58:41 crc kubenswrapper[4678]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 13 12:58:41 crc kubenswrapper[4678]: > Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.634085 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t844\" (UniqueName: \"kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844\") pod \"placement-96f2-account-create-4h6m4\" (UID: \"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4\") " pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.736089 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t844\" (UniqueName: \"kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844\") pod \"placement-96f2-account-create-4h6m4\" (UID: \"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4\") " pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.757171 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t844\" (UniqueName: \"kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844\") pod \"placement-96f2-account-create-4h6m4\" (UID: \"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4\") " pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.879015 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-j9fqj"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.879954 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.880239 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.882805 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6kwp6" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.882983 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.892961 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j9fqj"] Oct 13 12:58:41 crc kubenswrapper[4678]: I1013 12:58:41.933375 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5e8c-account-create-dhm4c"] Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.042195 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.042272 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2jff\" (UniqueName: \"kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.042308 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.042352 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.144187 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2jff\" (UniqueName: \"kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.144246 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.144271 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.144330 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.149566 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.154761 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.157661 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.163402 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2jff\" (UniqueName: \"kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff\") pod \"glance-db-sync-j9fqj\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.175276 4678 generic.go:334] "Generic (PLEG): container finished" podID="d624a6f3-db83-4d66-b86a-0d76348a9839" containerID="f938b2737eb47e152014cde3580556cad37961bdcea5df04ea7820970f20c5f5" exitCode=0 Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.175333 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e8c-account-create-dhm4c" event={"ID":"d624a6f3-db83-4d66-b86a-0d76348a9839","Type":"ContainerDied","Data":"f938b2737eb47e152014cde3580556cad37961bdcea5df04ea7820970f20c5f5"} Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.175357 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e8c-account-create-dhm4c" event={"ID":"d624a6f3-db83-4d66-b86a-0d76348a9839","Type":"ContainerStarted","Data":"7f44f075a96e3480eef41f8abb178dc9ef6bf897b05e964aa19225dc8fb10cf0"} Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.178828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"76d243e574655b96dd0ae76238a6ef99154d7521f11a51f0afb9ae97ac89ea24"} Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.317439 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-96f2-account-create-4h6m4"] Oct 13 12:58:42 crc kubenswrapper[4678]: I1013 12:58:42.320115 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j9fqj" Oct 13 12:58:42 crc kubenswrapper[4678]: W1013 12:58:42.567081 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefbce8cb_8fdf_4beb_b1c1_87937f5b7df4.slice/crio-92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e WatchSource:0}: Error finding container 92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e: Status 404 returned error can't find the container with id 92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.179893 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-j9fqj"] Oct 13 12:58:43 crc kubenswrapper[4678]: W1013 12:58:43.182943 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c40f306_6673_4695_86ed_32704f67754c.slice/crio-42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5 WatchSource:0}: Error finding container 42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5: Status 404 returned error can't find the container with id 42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5 Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.192376 4678 generic.go:334] "Generic (PLEG): container finished" podID="efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" containerID="49232af032db8d67bcac68322ae85a833a7502f74add7088130fed6d4ce902c3" exitCode=0 Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.192452 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96f2-account-create-4h6m4" event={"ID":"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4","Type":"ContainerDied","Data":"49232af032db8d67bcac68322ae85a833a7502f74add7088130fed6d4ce902c3"} Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.192479 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96f2-account-create-4h6m4" event={"ID":"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4","Type":"ContainerStarted","Data":"92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e"} Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.201432 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"72ad7e8b37922c82625435f2647f632078fd32b96e703e910d0518cec290e0c7"} Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.201473 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"706784b9a337f97c5ea8e45859b527f96de6a3cb9511327e705c3e7cbdfa0707"} Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.534504 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.670352 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpc5c\" (UniqueName: \"kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c\") pod \"d624a6f3-db83-4d66-b86a-0d76348a9839\" (UID: \"d624a6f3-db83-4d66-b86a-0d76348a9839\") " Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.689921 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c" (OuterVolumeSpecName: "kube-api-access-wpc5c") pod "d624a6f3-db83-4d66-b86a-0d76348a9839" (UID: "d624a6f3-db83-4d66-b86a-0d76348a9839"). InnerVolumeSpecName "kube-api-access-wpc5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:43 crc kubenswrapper[4678]: I1013 12:58:43.773827 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpc5c\" (UniqueName: \"kubernetes.io/projected/d624a6f3-db83-4d66-b86a-0d76348a9839-kube-api-access-wpc5c\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.218400 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"f59e615588b87c02abda39662a86afe12d1d003d892203b59b5e2729e6a0acd2"} Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.218454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"0b8bfa07d2456b9757668e5cc7395271f659a1123254c74078d1608f6c8f4372"} Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.222078 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5e8c-account-create-dhm4c" event={"ID":"d624a6f3-db83-4d66-b86a-0d76348a9839","Type":"ContainerDied","Data":"7f44f075a96e3480eef41f8abb178dc9ef6bf897b05e964aa19225dc8fb10cf0"} Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.222112 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f44f075a96e3480eef41f8abb178dc9ef6bf897b05e964aa19225dc8fb10cf0" Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.222177 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5e8c-account-create-dhm4c" Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.226904 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j9fqj" event={"ID":"4c40f306-6673-4695-86ed-32704f67754c","Type":"ContainerStarted","Data":"42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5"} Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.734817 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.788935 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t844\" (UniqueName: \"kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844\") pod \"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4\" (UID: \"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4\") " Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.796912 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844" (OuterVolumeSpecName: "kube-api-access-5t844") pod "efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" (UID: "efbce8cb-8fdf-4beb-b1c1-87937f5b7df4"). InnerVolumeSpecName "kube-api-access-5t844". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:44 crc kubenswrapper[4678]: I1013 12:58:44.890585 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t844\" (UniqueName: \"kubernetes.io/projected/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4-kube-api-access-5t844\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:45 crc kubenswrapper[4678]: I1013 12:58:45.239670 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-96f2-account-create-4h6m4" event={"ID":"efbce8cb-8fdf-4beb-b1c1-87937f5b7df4","Type":"ContainerDied","Data":"92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e"} Oct 13 12:58:45 crc kubenswrapper[4678]: I1013 12:58:45.239730 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d7e16715986a8f9ebfe88093660fec2709fe3938454a69b450b54e50ceca1e" Oct 13 12:58:45 crc kubenswrapper[4678]: I1013 12:58:45.239975 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-96f2-account-create-4h6m4" Oct 13 12:58:45 crc kubenswrapper[4678]: I1013 12:58:45.257076 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"eef836070912c49f88961b09415af3adbaebad456032448daab5a20576ea0530"} Oct 13 12:58:45 crc kubenswrapper[4678]: I1013 12:58:45.257118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"79aff96531d7142526ee4ef3be48ea20c5fdae846698829d077e8e9caab2d49c"} Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.275695 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"ec265f9f05d8dc7ea993fbbc3212bf8f65e1d04bb88a4c04aa1ebde36e5fa177"} Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.275963 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"17509cc7db9e4da0a7e02115c4a4fa6186e6ab91af0645efc63de3b8b937fc19"} Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.645583 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-dv2nq" podUID="2e724ee3-a6c7-4b75-9c60-3fe2f233b96b" containerName="ovn-controller" probeResult="failure" output=< Oct 13 12:58:46 crc kubenswrapper[4678]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 13 12:58:46 crc kubenswrapper[4678]: > Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.672386 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.674863 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4z6wg" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.907398 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dv2nq-config-gsknz"] Oct 13 12:58:46 crc kubenswrapper[4678]: E1013 12:58:46.909812 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.910070 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: E1013 12:58:46.910156 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d624a6f3-db83-4d66-b86a-0d76348a9839" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.910219 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d624a6f3-db83-4d66-b86a-0d76348a9839" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.910429 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d624a6f3-db83-4d66-b86a-0d76348a9839" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.910518 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" containerName="mariadb-account-create" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.911501 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.913889 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq-config-gsknz"] Oct 13 12:58:46 crc kubenswrapper[4678]: I1013 12:58:46.918418 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022753 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022794 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022839 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022871 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022925 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jnh4\" (UniqueName: \"kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.022952 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124267 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124337 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124370 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124423 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jnh4\" (UniqueName: \"kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124480 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124567 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124581 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.124647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.125273 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.126452 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.149780 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jnh4\" (UniqueName: \"kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4\") pod \"ovn-controller-dv2nq-config-gsknz\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.291349 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"6b9c8c020216b831db8a61c63ef5541a2f8caa8f1d14506d06c7b65d62d701e7"} Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.291390 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"612559e0bf4588dd29517d0aa862becf7f634a2ba1c21752ff7a985b58febdb2"} Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.291400 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"9c380dafb44ae54d9f0378bcf14c2c0d1e842f93be7dde78b1203428973457a6"} Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.303986 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.758261 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq-config-gsknz"] Oct 13 12:58:47 crc kubenswrapper[4678]: I1013 12:58:47.769660 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 13 12:58:47 crc kubenswrapper[4678]: W1013 12:58:47.781910 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod207bb3dc_2cb6_487d_bced_8b1424be18b8.slice/crio-6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47 WatchSource:0}: Error finding container 6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47: Status 404 returned error can't find the container with id 6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47 Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.092236 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.146173 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-7lkqp"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.147549 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.161863 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7lkqp"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.244701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb8c5\" (UniqueName: \"kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5\") pod \"barbican-db-create-7lkqp\" (UID: \"efdc002e-fb8c-4e54-8a8b-570e8b134cbc\") " pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.287555 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-k5zwg"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.289252 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.302487 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-k5zwg"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.330487 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-gsknz" event={"ID":"207bb3dc-2cb6-487d-bced-8b1424be18b8","Type":"ContainerStarted","Data":"2e7cdfe2be6d3789715e07cdb62d07d9714fabf3cfb95e224b3b886b7c1e720f"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.330527 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-gsknz" event={"ID":"207bb3dc-2cb6-487d-bced-8b1424be18b8","Type":"ContainerStarted","Data":"6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.343850 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"d8fb859d3583f40d6a4677cb214c94e8ecdae84c4c328c655caa579c34323862"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.343889 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"80bc53e66eabaf39a831e2c5efd54c6451dfcaa06883cced01ee198f8d0a4f0c"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.343898 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"678d3406b03a8af62deb6f77c75e7eae5c4cce7fa9b0989ea9f5d5c7d10440d0"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.343907 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e2337ebf-ba94-424a-9b45-fb68453bbf64","Type":"ContainerStarted","Data":"48d460ccfbf6309b7632ded5ac543ba2c783e8701f116fb9d0e93226ca457b8a"} Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.346294 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gvw8\" (UniqueName: \"kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8\") pod \"cinder-db-create-k5zwg\" (UID: \"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6\") " pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.346405 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb8c5\" (UniqueName: \"kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5\") pod \"barbican-db-create-7lkqp\" (UID: \"efdc002e-fb8c-4e54-8a8b-570e8b134cbc\") " pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.381907 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-dv2nq-config-gsknz" podStartSLOduration=2.381892346 podStartE2EDuration="2.381892346s" podCreationTimestamp="2025-10-13 12:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:48.378521998 +0000 UTC m=+916.463059882" watchObservedRunningTime="2025-10-13 12:58:48.381892346 +0000 UTC m=+916.466430220" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.391212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb8c5\" (UniqueName: \"kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5\") pod \"barbican-db-create-7lkqp\" (UID: \"efdc002e-fb8c-4e54-8a8b-570e8b134cbc\") " pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.451869 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gvw8\" (UniqueName: \"kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8\") pod \"cinder-db-create-k5zwg\" (UID: \"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6\") " pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.454771 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-b62wr"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.455571 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.534456356 podStartE2EDuration="25.455558275s" podCreationTimestamp="2025-10-13 12:58:23 +0000 UTC" firstStartedPulling="2025-10-13 12:58:41.556634739 +0000 UTC m=+909.641172623" lastFinishedPulling="2025-10-13 12:58:46.477736658 +0000 UTC m=+914.562274542" observedRunningTime="2025-10-13 12:58:48.439503085 +0000 UTC m=+916.524040989" watchObservedRunningTime="2025-10-13 12:58:48.455558275 +0000 UTC m=+916.540096159" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.456238 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.473875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gvw8\" (UniqueName: \"kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8\") pod \"cinder-db-create-k5zwg\" (UID: \"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6\") " pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.477514 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b62wr"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.529022 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.534425 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jmfrc"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.535468 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.538039 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xbmbf" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.538266 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.538300 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.538384 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.553174 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2dw\" (UniqueName: \"kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw\") pod \"neutron-db-create-b62wr\" (UID: \"93fd1326-7d69-497f-a770-ee1fb9de923d\") " pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.557130 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jmfrc"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.618074 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.654211 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.654558 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.654641 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2dw\" (UniqueName: \"kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw\") pod \"neutron-db-create-b62wr\" (UID: \"93fd1326-7d69-497f-a770-ee1fb9de923d\") " pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.654665 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjwj\" (UniqueName: \"kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.674490 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2dw\" (UniqueName: \"kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw\") pod \"neutron-db-create-b62wr\" (UID: \"93fd1326-7d69-497f-a770-ee1fb9de923d\") " pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.755710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.755754 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.755834 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hjwj\" (UniqueName: \"kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.762867 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.763718 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.769250 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.775538 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.776490 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.778471 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.785135 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hjwj\" (UniqueName: \"kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj\") pod \"keystone-db-sync-jmfrc\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.795210 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860659 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860711 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860739 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860819 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860853 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.860890 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2l7x\" (UniqueName: \"kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.864528 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.948617 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-k5zwg"] Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.965640 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.967073 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.967150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.968058 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.968195 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2l7x\" (UniqueName: \"kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.968404 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.968466 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.968523 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.973944 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.975412 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:48 crc kubenswrapper[4678]: I1013 12:58:48.977748 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.000023 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2l7x\" (UniqueName: \"kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x\") pod \"dnsmasq-dns-77585f5f8c-sg89w\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.083793 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-7lkqp"] Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.153952 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b62wr"] Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.187357 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jmfrc"] Oct 13 12:58:49 crc kubenswrapper[4678]: W1013 12:58:49.195228 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3931313a_c89a_442e_b955_56aeeaf0d200.slice/crio-a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7 WatchSource:0}: Error finding container a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7: Status 404 returned error can't find the container with id a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7 Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.261949 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.357018 4678 generic.go:334] "Generic (PLEG): container finished" podID="efdc002e-fb8c-4e54-8a8b-570e8b134cbc" containerID="701a66fd1f0a3896fed996332e51b2d169fc9d7b795b789a9deee5de1e1d7588" exitCode=0 Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.357121 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7lkqp" event={"ID":"efdc002e-fb8c-4e54-8a8b-570e8b134cbc","Type":"ContainerDied","Data":"701a66fd1f0a3896fed996332e51b2d169fc9d7b795b789a9deee5de1e1d7588"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.357145 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7lkqp" event={"ID":"efdc002e-fb8c-4e54-8a8b-570e8b134cbc","Type":"ContainerStarted","Data":"d658ed059b429dc4e7ecd5e157cae1aea02f9988ea8aa014aead946b4b219ce3"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.359366 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jmfrc" event={"ID":"3931313a-c89a-442e-b955-56aeeaf0d200","Type":"ContainerStarted","Data":"a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.360968 4678 generic.go:334] "Generic (PLEG): container finished" podID="1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" containerID="30f18b2f11249fb9e8355d2a7261d818e09322dcccd4a1534cee5d3130b2b37a" exitCode=0 Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.361002 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-k5zwg" event={"ID":"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6","Type":"ContainerDied","Data":"30f18b2f11249fb9e8355d2a7261d818e09322dcccd4a1534cee5d3130b2b37a"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.361016 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-k5zwg" event={"ID":"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6","Type":"ContainerStarted","Data":"01228e1fedd69e03af017ac9e1bac92c0964b25d7cdd8e78d0212345ade81dda"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.362072 4678 generic.go:334] "Generic (PLEG): container finished" podID="93fd1326-7d69-497f-a770-ee1fb9de923d" containerID="05972cb67214c4e5f83b5ee0949d02e462d179944ab6bf6a92c7908c4b2e5ba1" exitCode=0 Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.362113 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b62wr" event={"ID":"93fd1326-7d69-497f-a770-ee1fb9de923d","Type":"ContainerDied","Data":"05972cb67214c4e5f83b5ee0949d02e462d179944ab6bf6a92c7908c4b2e5ba1"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.362127 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b62wr" event={"ID":"93fd1326-7d69-497f-a770-ee1fb9de923d","Type":"ContainerStarted","Data":"5967b42f7770872df3c9086729e216a470fefdd727db444aeafc9b2517894064"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.363170 4678 generic.go:334] "Generic (PLEG): container finished" podID="207bb3dc-2cb6-487d-bced-8b1424be18b8" containerID="2e7cdfe2be6d3789715e07cdb62d07d9714fabf3cfb95e224b3b886b7c1e720f" exitCode=0 Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.364145 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-gsknz" event={"ID":"207bb3dc-2cb6-487d-bced-8b1424be18b8","Type":"ContainerDied","Data":"2e7cdfe2be6d3789715e07cdb62d07d9714fabf3cfb95e224b3b886b7c1e720f"} Oct 13 12:58:49 crc kubenswrapper[4678]: I1013 12:58:49.728967 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:58:49 crc kubenswrapper[4678]: W1013 12:58:49.739513 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf2eadc6_a42c_4029_a644_3efd3b585d88.slice/crio-3dc09dd2f31cf883cb75497ad60f402c5621b5d0a05d78cb8344bbbc7601c4ad WatchSource:0}: Error finding container 3dc09dd2f31cf883cb75497ad60f402c5621b5d0a05d78cb8344bbbc7601c4ad: Status 404 returned error can't find the container with id 3dc09dd2f31cf883cb75497ad60f402c5621b5d0a05d78cb8344bbbc7601c4ad Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.372207 4678 generic.go:334] "Generic (PLEG): container finished" podID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerID="b0762c9f3e8d151cff542a8ee34ae9ac467c1dccb24e346005a0ab1727940001" exitCode=0 Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.372565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" event={"ID":"bf2eadc6-a42c-4029-a644-3efd3b585d88","Type":"ContainerDied","Data":"b0762c9f3e8d151cff542a8ee34ae9ac467c1dccb24e346005a0ab1727940001"} Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.372595 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" event={"ID":"bf2eadc6-a42c-4029-a644-3efd3b585d88","Type":"ContainerStarted","Data":"3dc09dd2f31cf883cb75497ad60f402c5621b5d0a05d78cb8344bbbc7601c4ad"} Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.676032 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806549 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806623 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806677 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jnh4\" (UniqueName: \"kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806720 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806772 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.806828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn\") pod \"207bb3dc-2cb6-487d-bced-8b1424be18b8\" (UID: \"207bb3dc-2cb6-487d-bced-8b1424be18b8\") " Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.807169 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.807200 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.808719 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts" (OuterVolumeSpecName: "scripts") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.808772 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run" (OuterVolumeSpecName: "var-run") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.809511 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.811504 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4" (OuterVolumeSpecName: "kube-api-access-4jnh4") pod "207bb3dc-2cb6-487d-bced-8b1424be18b8" (UID: "207bb3dc-2cb6-487d-bced-8b1424be18b8"). InnerVolumeSpecName "kube-api-access-4jnh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908566 4678 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908596 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908607 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jnh4\" (UniqueName: \"kubernetes.io/projected/207bb3dc-2cb6-487d-bced-8b1424be18b8-kube-api-access-4jnh4\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908616 4678 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908625 4678 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/207bb3dc-2cb6-487d-bced-8b1424be18b8-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.908633 4678 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/207bb3dc-2cb6-487d-bced-8b1424be18b8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.954445 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.955319 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:50 crc kubenswrapper[4678]: I1013 12:58:50.956759 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.010156 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gvw8\" (UniqueName: \"kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8\") pod \"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6\" (UID: \"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6\") " Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.010199 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c2dw\" (UniqueName: \"kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw\") pod \"93fd1326-7d69-497f-a770-ee1fb9de923d\" (UID: \"93fd1326-7d69-497f-a770-ee1fb9de923d\") " Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.010316 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb8c5\" (UniqueName: \"kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5\") pod \"efdc002e-fb8c-4e54-8a8b-570e8b134cbc\" (UID: \"efdc002e-fb8c-4e54-8a8b-570e8b134cbc\") " Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.021507 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5" (OuterVolumeSpecName: "kube-api-access-xb8c5") pod "efdc002e-fb8c-4e54-8a8b-570e8b134cbc" (UID: "efdc002e-fb8c-4e54-8a8b-570e8b134cbc"). InnerVolumeSpecName "kube-api-access-xb8c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.021600 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8" (OuterVolumeSpecName: "kube-api-access-5gvw8") pod "1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" (UID: "1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6"). InnerVolumeSpecName "kube-api-access-5gvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.037211 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw" (OuterVolumeSpecName: "kube-api-access-5c2dw") pod "93fd1326-7d69-497f-a770-ee1fb9de923d" (UID: "93fd1326-7d69-497f-a770-ee1fb9de923d"). InnerVolumeSpecName "kube-api-access-5c2dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.112402 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gvw8\" (UniqueName: \"kubernetes.io/projected/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6-kube-api-access-5gvw8\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.112428 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c2dw\" (UniqueName: \"kubernetes.io/projected/93fd1326-7d69-497f-a770-ee1fb9de923d-kube-api-access-5c2dw\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.112437 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb8c5\" (UniqueName: \"kubernetes.io/projected/efdc002e-fb8c-4e54-8a8b-570e8b134cbc-kube-api-access-xb8c5\") on node \"crc\" DevicePath \"\"" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.382474 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-k5zwg" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.382514 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-k5zwg" event={"ID":"1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6","Type":"ContainerDied","Data":"01228e1fedd69e03af017ac9e1bac92c0964b25d7cdd8e78d0212345ade81dda"} Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.382860 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01228e1fedd69e03af017ac9e1bac92c0964b25d7cdd8e78d0212345ade81dda" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.384745 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b62wr" event={"ID":"93fd1326-7d69-497f-a770-ee1fb9de923d","Type":"ContainerDied","Data":"5967b42f7770872df3c9086729e216a470fefdd727db444aeafc9b2517894064"} Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.384780 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5967b42f7770872df3c9086729e216a470fefdd727db444aeafc9b2517894064" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.384787 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b62wr" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.386902 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" event={"ID":"bf2eadc6-a42c-4029-a644-3efd3b585d88","Type":"ContainerStarted","Data":"2b39cfde21a792b1382e0e78dddc89f9877ffec87f645cbe7f81f7512b533f9e"} Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.387025 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.389139 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-gsknz" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.389150 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-gsknz" event={"ID":"207bb3dc-2cb6-487d-bced-8b1424be18b8","Type":"ContainerDied","Data":"6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47"} Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.389178 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6069581b4e035c4c77cae34caea721cd4cf87a5f138883204ff2ec0ef4e99a47" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.391045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-7lkqp" event={"ID":"efdc002e-fb8c-4e54-8a8b-570e8b134cbc","Type":"ContainerDied","Data":"d658ed059b429dc4e7ecd5e157cae1aea02f9988ea8aa014aead946b4b219ce3"} Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.391090 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-7lkqp" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.391100 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d658ed059b429dc4e7ecd5e157cae1aea02f9988ea8aa014aead946b4b219ce3" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.426676 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" podStartSLOduration=3.426658135 podStartE2EDuration="3.426658135s" podCreationTimestamp="2025-10-13 12:58:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:58:51.41385663 +0000 UTC m=+919.498394514" watchObservedRunningTime="2025-10-13 12:58:51.426658135 +0000 UTC m=+919.511196019" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.456214 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-dv2nq-config-gsknz"] Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.463089 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-dv2nq-config-gsknz"] Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575196 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-dv2nq-config-68jx2"] Oct 13 12:58:51 crc kubenswrapper[4678]: E1013 12:58:51.575590 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207bb3dc-2cb6-487d-bced-8b1424be18b8" containerName="ovn-config" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575611 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="207bb3dc-2cb6-487d-bced-8b1424be18b8" containerName="ovn-config" Oct 13 12:58:51 crc kubenswrapper[4678]: E1013 12:58:51.575625 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efdc002e-fb8c-4e54-8a8b-570e8b134cbc" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575634 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="efdc002e-fb8c-4e54-8a8b-570e8b134cbc" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: E1013 12:58:51.575662 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575670 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: E1013 12:58:51.575684 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93fd1326-7d69-497f-a770-ee1fb9de923d" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575691 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="93fd1326-7d69-497f-a770-ee1fb9de923d" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575871 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575912 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="93fd1326-7d69-497f-a770-ee1fb9de923d" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575933 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="207bb3dc-2cb6-487d-bced-8b1424be18b8" containerName="ovn-config" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.575947 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="efdc002e-fb8c-4e54-8a8b-570e8b134cbc" containerName="mariadb-database-create" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.576591 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.578272 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.587002 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq-config-68jx2"] Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.636419 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-dv2nq" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727723 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjtlg\" (UniqueName: \"kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727805 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727856 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727885 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727903 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.727932 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.829767 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.829812 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.829843 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.829943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjtlg\" (UniqueName: \"kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.829986 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.830040 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.830161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.830191 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.830286 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.830791 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.832036 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.859961 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjtlg\" (UniqueName: \"kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg\") pod \"ovn-controller-dv2nq-config-68jx2\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:51 crc kubenswrapper[4678]: I1013 12:58:51.890306 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:58:52 crc kubenswrapper[4678]: I1013 12:58:52.607883 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207bb3dc-2cb6-487d-bced-8b1424be18b8" path="/var/lib/kubelet/pods/207bb3dc-2cb6-487d-bced-8b1424be18b8/volumes" Oct 13 12:58:55 crc kubenswrapper[4678]: I1013 12:58:55.506000 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 12:58:55 crc kubenswrapper[4678]: I1013 12:58:55.506362 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 12:58:55 crc kubenswrapper[4678]: I1013 12:58:55.506441 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 12:58:55 crc kubenswrapper[4678]: I1013 12:58:55.507159 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 12:58:55 crc kubenswrapper[4678]: I1013 12:58:55.507224 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470" gracePeriod=600 Oct 13 12:58:56 crc kubenswrapper[4678]: I1013 12:58:56.444534 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470" exitCode=0 Oct 13 12:58:56 crc kubenswrapper[4678]: I1013 12:58:56.444597 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470"} Oct 13 12:58:56 crc kubenswrapper[4678]: I1013 12:58:56.444900 4678 scope.go:117] "RemoveContainer" containerID="eb2457d967e2096f441307642b161d4ff7c1a04040b8d8c1d768d14a0e9bb002" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.175196 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-101c-account-create-n6vrn"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.177560 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.179237 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.233655 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-101c-account-create-n6vrn"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.259965 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzplk\" (UniqueName: \"kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk\") pod \"barbican-101c-account-create-n6vrn\" (UID: \"6dbff364-1cc3-44d1-af82-63bbbc39d2e3\") " pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.273174 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-06d9-account-create-tjsv6"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.274590 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.276603 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.287493 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-06d9-account-create-tjsv6"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.362270 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzplk\" (UniqueName: \"kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk\") pod \"barbican-101c-account-create-n6vrn\" (UID: \"6dbff364-1cc3-44d1-af82-63bbbc39d2e3\") " pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.362463 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw5mb\" (UniqueName: \"kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb\") pod \"cinder-06d9-account-create-tjsv6\" (UID: \"0d1f30c5-e273-4bd7-9d62-62f26b253fca\") " pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.384583 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzplk\" (UniqueName: \"kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk\") pod \"barbican-101c-account-create-n6vrn\" (UID: \"6dbff364-1cc3-44d1-af82-63bbbc39d2e3\") " pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.463766 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw5mb\" (UniqueName: \"kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb\") pod \"cinder-06d9-account-create-tjsv6\" (UID: \"0d1f30c5-e273-4bd7-9d62-62f26b253fca\") " pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.482601 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw5mb\" (UniqueName: \"kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb\") pod \"cinder-06d9-account-create-tjsv6\" (UID: \"0d1f30c5-e273-4bd7-9d62-62f26b253fca\") " pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.530626 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.567706 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-3012-account-create-bhsxw"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.568763 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.570685 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.577631 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3012-account-create-bhsxw"] Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.598307 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.667300 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7hq5\" (UniqueName: \"kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5\") pod \"neutron-3012-account-create-bhsxw\" (UID: \"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce\") " pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.769380 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7hq5\" (UniqueName: \"kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5\") pod \"neutron-3012-account-create-bhsxw\" (UID: \"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce\") " pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.784925 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7hq5\" (UniqueName: \"kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5\") pod \"neutron-3012-account-create-bhsxw\" (UID: \"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce\") " pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:58:58 crc kubenswrapper[4678]: I1013 12:58:58.883333 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:58:59 crc kubenswrapper[4678]: I1013 12:58:59.263238 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:58:59 crc kubenswrapper[4678]: I1013 12:58:59.330859 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:58:59 crc kubenswrapper[4678]: I1013 12:58:59.331093 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-2clp7" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="dnsmasq-dns" containerID="cri-o://85f9bfa30f288b130591f59db2a145fa3af223ee4eed2591ca3d0c91d2f2bad2" gracePeriod=10 Oct 13 12:59:00 crc kubenswrapper[4678]: I1013 12:59:00.478082 4678 generic.go:334] "Generic (PLEG): container finished" podID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerID="85f9bfa30f288b130591f59db2a145fa3af223ee4eed2591ca3d0c91d2f2bad2" exitCode=0 Oct 13 12:59:00 crc kubenswrapper[4678]: I1013 12:59:00.478165 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2clp7" event={"ID":"f5258b79-3ee1-4989-bc3c-23711b14f1a6","Type":"ContainerDied","Data":"85f9bfa30f288b130591f59db2a145fa3af223ee4eed2591ca3d0c91d2f2bad2"} Oct 13 12:59:01 crc kubenswrapper[4678]: E1013 12:59:01.232247 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone:current-podified" Oct 13 12:59:01 crc kubenswrapper[4678]: E1013 12:59:01.232500 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6hjwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-jmfrc_openstack(3931313a-c89a-442e-b955-56aeeaf0d200): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:59:01 crc kubenswrapper[4678]: E1013 12:59:01.233785 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-jmfrc" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" Oct 13 12:59:01 crc kubenswrapper[4678]: E1013 12:59:01.507226 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone:current-podified\\\"\"" pod="openstack/keystone-db-sync-jmfrc" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.595611 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:59:01 crc kubenswrapper[4678]: W1013 12:59:01.641965 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6f90349_cc47_4b03_85fc_d898d54cd69d.slice/crio-962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b WatchSource:0}: Error finding container 962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b: Status 404 returned error can't find the container with id 962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.643986 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-dv2nq-config-68jx2"] Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.721620 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb\") pod \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.721730 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr7h6\" (UniqueName: \"kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6\") pod \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.721767 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config\") pod \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.721828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc\") pod \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.721877 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb\") pod \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\" (UID: \"f5258b79-3ee1-4989-bc3c-23711b14f1a6\") " Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.732000 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6" (OuterVolumeSpecName: "kube-api-access-gr7h6") pod "f5258b79-3ee1-4989-bc3c-23711b14f1a6" (UID: "f5258b79-3ee1-4989-bc3c-23711b14f1a6"). InnerVolumeSpecName "kube-api-access-gr7h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.778779 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f5258b79-3ee1-4989-bc3c-23711b14f1a6" (UID: "f5258b79-3ee1-4989-bc3c-23711b14f1a6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.788572 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config" (OuterVolumeSpecName: "config") pod "f5258b79-3ee1-4989-bc3c-23711b14f1a6" (UID: "f5258b79-3ee1-4989-bc3c-23711b14f1a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.804064 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f5258b79-3ee1-4989-bc3c-23711b14f1a6" (UID: "f5258b79-3ee1-4989-bc3c-23711b14f1a6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.811074 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f5258b79-3ee1-4989-bc3c-23711b14f1a6" (UID: "f5258b79-3ee1-4989-bc3c-23711b14f1a6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.823290 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr7h6\" (UniqueName: \"kubernetes.io/projected/f5258b79-3ee1-4989-bc3c-23711b14f1a6-kube-api-access-gr7h6\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.823316 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.823327 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.823337 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.823346 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f5258b79-3ee1-4989-bc3c-23711b14f1a6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.912274 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-3012-account-create-bhsxw"] Oct 13 12:59:01 crc kubenswrapper[4678]: W1013 12:59:01.921030 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9182c1b1_e2f1_48b6_85ec_459d42ccc2ce.slice/crio-126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d WatchSource:0}: Error finding container 126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d: Status 404 returned error can't find the container with id 126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d Oct 13 12:59:01 crc kubenswrapper[4678]: I1013 12:59:01.995028 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-06d9-account-create-tjsv6"] Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.002629 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-101c-account-create-n6vrn"] Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.516549 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2clp7" event={"ID":"f5258b79-3ee1-4989-bc3c-23711b14f1a6","Type":"ContainerDied","Data":"c2453a2760ceb1998afb914eb5c431fa647e06b46f80c1f4dd93b53c5996e0a4"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.516601 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2clp7" Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.516838 4678 scope.go:117] "RemoveContainer" containerID="85f9bfa30f288b130591f59db2a145fa3af223ee4eed2591ca3d0c91d2f2bad2" Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.521907 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6f90349-cc47-4b03-85fc-d898d54cd69d" containerID="0f6deaacf2953924ac56461237f19f9d5d0f1a8c394a76f29e0405508a446019" exitCode=0 Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.521987 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-68jx2" event={"ID":"a6f90349-cc47-4b03-85fc-d898d54cd69d","Type":"ContainerDied","Data":"0f6deaacf2953924ac56461237f19f9d5d0f1a8c394a76f29e0405508a446019"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.522017 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-68jx2" event={"ID":"a6f90349-cc47-4b03-85fc-d898d54cd69d","Type":"ContainerStarted","Data":"962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.524931 4678 generic.go:334] "Generic (PLEG): container finished" podID="6dbff364-1cc3-44d1-af82-63bbbc39d2e3" containerID="6635145ecad9eb42dc99dd83ebfd65d3a214560efe747c3a5f49ea8354df55a3" exitCode=0 Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.524983 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-101c-account-create-n6vrn" event={"ID":"6dbff364-1cc3-44d1-af82-63bbbc39d2e3","Type":"ContainerDied","Data":"6635145ecad9eb42dc99dd83ebfd65d3a214560efe747c3a5f49ea8354df55a3"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.525003 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-101c-account-create-n6vrn" event={"ID":"6dbff364-1cc3-44d1-af82-63bbbc39d2e3","Type":"ContainerStarted","Data":"28ef9adfa264594eaf3ccd0dc4b71c18b01c493060b153586978c29c759cb36e"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.527160 4678 generic.go:334] "Generic (PLEG): container finished" podID="9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" containerID="f64fa91748745d96b80fa05e4d45f8b497f118f1943296f5bd98b13a9929cd4a" exitCode=0 Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.527210 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3012-account-create-bhsxw" event={"ID":"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce","Type":"ContainerDied","Data":"f64fa91748745d96b80fa05e4d45f8b497f118f1943296f5bd98b13a9929cd4a"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.527228 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3012-account-create-bhsxw" event={"ID":"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce","Type":"ContainerStarted","Data":"126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.529270 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.532953 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j9fqj" event={"ID":"4c40f306-6673-4695-86ed-32704f67754c","Type":"ContainerStarted","Data":"23f30c99c952ff26c8254f54695d722b297d62ee3cc21813ba5a2e987e238eb4"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.537423 4678 generic.go:334] "Generic (PLEG): container finished" podID="0d1f30c5-e273-4bd7-9d62-62f26b253fca" containerID="9a5fdeed6f438cf5638f1d8b942d9e37b0926110d365a83f936438e2cf76aa36" exitCode=0 Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.537488 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-06d9-account-create-tjsv6" event={"ID":"0d1f30c5-e273-4bd7-9d62-62f26b253fca","Type":"ContainerDied","Data":"9a5fdeed6f438cf5638f1d8b942d9e37b0926110d365a83f936438e2cf76aa36"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.537527 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-06d9-account-create-tjsv6" event={"ID":"0d1f30c5-e273-4bd7-9d62-62f26b253fca","Type":"ContainerStarted","Data":"9e0878ad7f3e2dfc72b51ff5b13cd8c0337dd3ff4eadc53c8fa5fc035f30339a"} Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.545201 4678 scope.go:117] "RemoveContainer" containerID="3c6511b004faf36fb28e55e449c0f68595773a33da06e995c4976439434c5085" Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.649578 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.664796 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2clp7"] Oct 13 12:59:02 crc kubenswrapper[4678]: I1013 12:59:02.673143 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-j9fqj" podStartSLOduration=3.433030859 podStartE2EDuration="21.673115889s" podCreationTimestamp="2025-10-13 12:58:41 +0000 UTC" firstStartedPulling="2025-10-13 12:58:43.18523169 +0000 UTC m=+911.269769594" lastFinishedPulling="2025-10-13 12:59:01.42531673 +0000 UTC m=+929.509854624" observedRunningTime="2025-10-13 12:59:02.66131952 +0000 UTC m=+930.745857414" watchObservedRunningTime="2025-10-13 12:59:02.673115889 +0000 UTC m=+930.757653803" Oct 13 12:59:03 crc kubenswrapper[4678]: I1013 12:59:03.932763 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.069877 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.069941 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.070041 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjtlg\" (UniqueName: \"kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.070141 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.070203 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.070257 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn\") pod \"a6f90349-cc47-4b03-85fc-d898d54cd69d\" (UID: \"a6f90349-cc47-4b03-85fc-d898d54cd69d\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.070637 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.071231 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run" (OuterVolumeSpecName: "var-run") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.071287 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.071651 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.072084 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts" (OuterVolumeSpecName: "scripts") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.075589 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg" (OuterVolumeSpecName: "kube-api-access-vjtlg") pod "a6f90349-cc47-4b03-85fc-d898d54cd69d" (UID: "a6f90349-cc47-4b03-85fc-d898d54cd69d"). InnerVolumeSpecName "kube-api-access-vjtlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.136670 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.142754 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.150025 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172633 4678 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172673 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6f90349-cc47-4b03-85fc-d898d54cd69d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172687 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjtlg\" (UniqueName: \"kubernetes.io/projected/a6f90349-cc47-4b03-85fc-d898d54cd69d-kube-api-access-vjtlg\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172702 4678 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172715 4678 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.172727 4678 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6f90349-cc47-4b03-85fc-d898d54cd69d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.273227 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7hq5\" (UniqueName: \"kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5\") pod \"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce\" (UID: \"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.273336 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzplk\" (UniqueName: \"kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk\") pod \"6dbff364-1cc3-44d1-af82-63bbbc39d2e3\" (UID: \"6dbff364-1cc3-44d1-af82-63bbbc39d2e3\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.273377 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw5mb\" (UniqueName: \"kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb\") pod \"0d1f30c5-e273-4bd7-9d62-62f26b253fca\" (UID: \"0d1f30c5-e273-4bd7-9d62-62f26b253fca\") " Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.278609 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5" (OuterVolumeSpecName: "kube-api-access-v7hq5") pod "9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" (UID: "9182c1b1-e2f1-48b6-85ec-459d42ccc2ce"). InnerVolumeSpecName "kube-api-access-v7hq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.279825 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk" (OuterVolumeSpecName: "kube-api-access-kzplk") pod "6dbff364-1cc3-44d1-af82-63bbbc39d2e3" (UID: "6dbff364-1cc3-44d1-af82-63bbbc39d2e3"). InnerVolumeSpecName "kube-api-access-kzplk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.279874 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb" (OuterVolumeSpecName: "kube-api-access-sw5mb") pod "0d1f30c5-e273-4bd7-9d62-62f26b253fca" (UID: "0d1f30c5-e273-4bd7-9d62-62f26b253fca"). InnerVolumeSpecName "kube-api-access-sw5mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.375231 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7hq5\" (UniqueName: \"kubernetes.io/projected/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce-kube-api-access-v7hq5\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.375264 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzplk\" (UniqueName: \"kubernetes.io/projected/6dbff364-1cc3-44d1-af82-63bbbc39d2e3-kube-api-access-kzplk\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.375273 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw5mb\" (UniqueName: \"kubernetes.io/projected/0d1f30c5-e273-4bd7-9d62-62f26b253fca-kube-api-access-sw5mb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.563804 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-06d9-account-create-tjsv6" event={"ID":"0d1f30c5-e273-4bd7-9d62-62f26b253fca","Type":"ContainerDied","Data":"9e0878ad7f3e2dfc72b51ff5b13cd8c0337dd3ff4eadc53c8fa5fc035f30339a"} Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.563850 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-06d9-account-create-tjsv6" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.563869 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e0878ad7f3e2dfc72b51ff5b13cd8c0337dd3ff4eadc53c8fa5fc035f30339a" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.565802 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-dv2nq-config-68jx2" event={"ID":"a6f90349-cc47-4b03-85fc-d898d54cd69d","Type":"ContainerDied","Data":"962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b"} Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.565839 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="962c6861f41499a59895cecb1fcf1e2dfbe4823420a39140abdfa16dbad7fa6b" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.566099 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-dv2nq-config-68jx2" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.567060 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-101c-account-create-n6vrn" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.567108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-101c-account-create-n6vrn" event={"ID":"6dbff364-1cc3-44d1-af82-63bbbc39d2e3","Type":"ContainerDied","Data":"28ef9adfa264594eaf3ccd0dc4b71c18b01c493060b153586978c29c759cb36e"} Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.567175 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ef9adfa264594eaf3ccd0dc4b71c18b01c493060b153586978c29c759cb36e" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.568420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-3012-account-create-bhsxw" event={"ID":"9182c1b1-e2f1-48b6-85ec-459d42ccc2ce","Type":"ContainerDied","Data":"126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d"} Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.568447 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="126b0beca644c39569bf2a23d81af1e579a95c9311036110116deb938f849e2d" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.568489 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-3012-account-create-bhsxw" Oct 13 12:59:04 crc kubenswrapper[4678]: I1013 12:59:04.609182 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" path="/var/lib/kubelet/pods/f5258b79-3ee1-4989-bc3c-23711b14f1a6/volumes" Oct 13 12:59:05 crc kubenswrapper[4678]: I1013 12:59:05.083876 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-dv2nq-config-68jx2"] Oct 13 12:59:05 crc kubenswrapper[4678]: I1013 12:59:05.095810 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-dv2nq-config-68jx2"] Oct 13 12:59:06 crc kubenswrapper[4678]: I1013 12:59:06.602138 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6f90349-cc47-4b03-85fc-d898d54cd69d" path="/var/lib/kubelet/pods/a6f90349-cc47-4b03-85fc-d898d54cd69d/volumes" Oct 13 12:59:10 crc kubenswrapper[4678]: I1013 12:59:10.625805 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j9fqj" event={"ID":"4c40f306-6673-4695-86ed-32704f67754c","Type":"ContainerDied","Data":"23f30c99c952ff26c8254f54695d722b297d62ee3cc21813ba5a2e987e238eb4"} Oct 13 12:59:10 crc kubenswrapper[4678]: I1013 12:59:10.625730 4678 generic.go:334] "Generic (PLEG): container finished" podID="4c40f306-6673-4695-86ed-32704f67754c" containerID="23f30c99c952ff26c8254f54695d722b297d62ee3cc21813ba5a2e987e238eb4" exitCode=0 Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.022603 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j9fqj" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.163478 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2jff\" (UniqueName: \"kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff\") pod \"4c40f306-6673-4695-86ed-32704f67754c\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.163585 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle\") pod \"4c40f306-6673-4695-86ed-32704f67754c\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.163612 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data\") pod \"4c40f306-6673-4695-86ed-32704f67754c\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.163660 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data\") pod \"4c40f306-6673-4695-86ed-32704f67754c\" (UID: \"4c40f306-6673-4695-86ed-32704f67754c\") " Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.175337 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4c40f306-6673-4695-86ed-32704f67754c" (UID: "4c40f306-6673-4695-86ed-32704f67754c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.176329 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff" (OuterVolumeSpecName: "kube-api-access-q2jff") pod "4c40f306-6673-4695-86ed-32704f67754c" (UID: "4c40f306-6673-4695-86ed-32704f67754c"). InnerVolumeSpecName "kube-api-access-q2jff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.186862 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c40f306-6673-4695-86ed-32704f67754c" (UID: "4c40f306-6673-4695-86ed-32704f67754c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.227664 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data" (OuterVolumeSpecName: "config-data") pod "4c40f306-6673-4695-86ed-32704f67754c" (UID: "4c40f306-6673-4695-86ed-32704f67754c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.266080 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2jff\" (UniqueName: \"kubernetes.io/projected/4c40f306-6673-4695-86ed-32704f67754c-kube-api-access-q2jff\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.266105 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.266116 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.266126 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c40f306-6673-4695-86ed-32704f67754c-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.646919 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-j9fqj" event={"ID":"4c40f306-6673-4695-86ed-32704f67754c","Type":"ContainerDied","Data":"42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5"} Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.646960 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42868d10fff8d7479f79ff265ddfba8b3ccefa5ff09c195d39537622ada4c8f5" Oct 13 12:59:12 crc kubenswrapper[4678]: I1013 12:59:12.647030 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-j9fqj" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.053981 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055246 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="dnsmasq-dns" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055317 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="dnsmasq-dns" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055374 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055436 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055491 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f90349-cc47-4b03-85fc-d898d54cd69d" containerName="ovn-config" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055551 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f90349-cc47-4b03-85fc-d898d54cd69d" containerName="ovn-config" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055705 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="init" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055762 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="init" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055819 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1f30c5-e273-4bd7-9d62-62f26b253fca" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055876 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1f30c5-e273-4bd7-9d62-62f26b253fca" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.055933 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c40f306-6673-4695-86ed-32704f67754c" containerName="glance-db-sync" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.055995 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c40f306-6673-4695-86ed-32704f67754c" containerName="glance-db-sync" Oct 13 12:59:13 crc kubenswrapper[4678]: E1013 12:59:13.056069 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dbff364-1cc3-44d1-af82-63bbbc39d2e3" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056133 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dbff364-1cc3-44d1-af82-63bbbc39d2e3" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056319 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dbff364-1cc3-44d1-af82-63bbbc39d2e3" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056385 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d1f30c5-e273-4bd7-9d62-62f26b253fca" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056441 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5258b79-3ee1-4989-bc3c-23711b14f1a6" containerName="dnsmasq-dns" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056506 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c40f306-6673-4695-86ed-32704f67754c" containerName="glance-db-sync" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056565 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f90349-cc47-4b03-85fc-d898d54cd69d" containerName="ovn-config" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.056645 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" containerName="mariadb-account-create" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.057698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.080310 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182482 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182563 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182589 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9smh4\" (UniqueName: \"kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182778 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.182890 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284409 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9smh4\" (UniqueName: \"kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284475 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284498 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284520 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.284555 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.285378 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.285392 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.285960 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.286180 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.286498 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.307653 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9smh4\" (UniqueName: \"kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4\") pod \"dnsmasq-dns-7ff5475cc9-zm25r\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.379843 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:13 crc kubenswrapper[4678]: I1013 12:59:13.818696 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:13 crc kubenswrapper[4678]: W1013 12:59:13.819245 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod777f277f_fdea_4b55_bcbf_d8f6b7419656.slice/crio-3384ae57541ed7f3a9e76477383d274b4a5227b8da04ec279b582e10c6119616 WatchSource:0}: Error finding container 3384ae57541ed7f3a9e76477383d274b4a5227b8da04ec279b582e10c6119616: Status 404 returned error can't find the container with id 3384ae57541ed7f3a9e76477383d274b4a5227b8da04ec279b582e10c6119616 Oct 13 12:59:14 crc kubenswrapper[4678]: I1013 12:59:14.664302 4678 generic.go:334] "Generic (PLEG): container finished" podID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerID="e618c544475fd80f774a4aa8a1539945609241fc1c8cf2b3b1abc218d3522a2f" exitCode=0 Oct 13 12:59:14 crc kubenswrapper[4678]: I1013 12:59:14.664373 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" event={"ID":"777f277f-fdea-4b55-bcbf-d8f6b7419656","Type":"ContainerDied","Data":"e618c544475fd80f774a4aa8a1539945609241fc1c8cf2b3b1abc218d3522a2f"} Oct 13 12:59:14 crc kubenswrapper[4678]: I1013 12:59:14.664586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" event={"ID":"777f277f-fdea-4b55-bcbf-d8f6b7419656","Type":"ContainerStarted","Data":"3384ae57541ed7f3a9e76477383d274b4a5227b8da04ec279b582e10c6119616"} Oct 13 12:59:15 crc kubenswrapper[4678]: I1013 12:59:15.674287 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" event={"ID":"777f277f-fdea-4b55-bcbf-d8f6b7419656","Type":"ContainerStarted","Data":"164846cd982fa5935ac9132f26e56706fb81a779b26392cf73a44afe6b752fd2"} Oct 13 12:59:15 crc kubenswrapper[4678]: I1013 12:59:15.674711 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:15 crc kubenswrapper[4678]: I1013 12:59:15.690498 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" podStartSLOduration=2.6904803680000002 podStartE2EDuration="2.690480368s" podCreationTimestamp="2025-10-13 12:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:15.688403494 +0000 UTC m=+943.772941408" watchObservedRunningTime="2025-10-13 12:59:15.690480368 +0000 UTC m=+943.775018252" Oct 13 12:59:18 crc kubenswrapper[4678]: I1013 12:59:18.716126 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jmfrc" event={"ID":"3931313a-c89a-442e-b955-56aeeaf0d200","Type":"ContainerStarted","Data":"a21d4cfbd1158a33b8fcbd4c06b615d24c72654d420145c0013dd93e67aed367"} Oct 13 12:59:18 crc kubenswrapper[4678]: I1013 12:59:18.762438 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jmfrc" podStartSLOduration=2.483368711 podStartE2EDuration="30.762407199s" podCreationTimestamp="2025-10-13 12:58:48 +0000 UTC" firstStartedPulling="2025-10-13 12:58:49.198642986 +0000 UTC m=+917.283180870" lastFinishedPulling="2025-10-13 12:59:17.477681464 +0000 UTC m=+945.562219358" observedRunningTime="2025-10-13 12:59:18.74069641 +0000 UTC m=+946.825234334" watchObservedRunningTime="2025-10-13 12:59:18.762407199 +0000 UTC m=+946.846945123" Oct 13 12:59:20 crc kubenswrapper[4678]: I1013 12:59:20.741043 4678 generic.go:334] "Generic (PLEG): container finished" podID="3931313a-c89a-442e-b955-56aeeaf0d200" containerID="a21d4cfbd1158a33b8fcbd4c06b615d24c72654d420145c0013dd93e67aed367" exitCode=0 Oct 13 12:59:20 crc kubenswrapper[4678]: I1013 12:59:20.741180 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jmfrc" event={"ID":"3931313a-c89a-442e-b955-56aeeaf0d200","Type":"ContainerDied","Data":"a21d4cfbd1158a33b8fcbd4c06b615d24c72654d420145c0013dd93e67aed367"} Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.144525 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.279020 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle\") pod \"3931313a-c89a-442e-b955-56aeeaf0d200\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.279076 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hjwj\" (UniqueName: \"kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj\") pod \"3931313a-c89a-442e-b955-56aeeaf0d200\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.279291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data\") pod \"3931313a-c89a-442e-b955-56aeeaf0d200\" (UID: \"3931313a-c89a-442e-b955-56aeeaf0d200\") " Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.288850 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj" (OuterVolumeSpecName: "kube-api-access-6hjwj") pod "3931313a-c89a-442e-b955-56aeeaf0d200" (UID: "3931313a-c89a-442e-b955-56aeeaf0d200"). InnerVolumeSpecName "kube-api-access-6hjwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.304769 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3931313a-c89a-442e-b955-56aeeaf0d200" (UID: "3931313a-c89a-442e-b955-56aeeaf0d200"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.323782 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data" (OuterVolumeSpecName: "config-data") pod "3931313a-c89a-442e-b955-56aeeaf0d200" (UID: "3931313a-c89a-442e-b955-56aeeaf0d200"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.381824 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.381861 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3931313a-c89a-442e-b955-56aeeaf0d200-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.381875 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hjwj\" (UniqueName: \"kubernetes.io/projected/3931313a-c89a-442e-b955-56aeeaf0d200-kube-api-access-6hjwj\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.770823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jmfrc" event={"ID":"3931313a-c89a-442e-b955-56aeeaf0d200","Type":"ContainerDied","Data":"a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7"} Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.770889 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3beaaf153bd25167abf43911406e282d031e38e53d42e9ebf595e830c73a4d7" Oct 13 12:59:22 crc kubenswrapper[4678]: I1013 12:59:22.770908 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jmfrc" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.165888 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.166211 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="dnsmasq-dns" containerID="cri-o://164846cd982fa5935ac9132f26e56706fb81a779b26392cf73a44afe6b752fd2" gracePeriod=10 Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.167191 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.176592 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-q58ds"] Oct 13 12:59:23 crc kubenswrapper[4678]: E1013 12:59:23.177100 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" containerName="keystone-db-sync" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.177150 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" containerName="keystone-db-sync" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.177366 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" containerName="keystone-db-sync" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.177862 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.187295 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.187476 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.187865 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xbmbf" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.188158 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.204154 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q58ds"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.230426 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.231676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.275701 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302260 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302318 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302346 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302387 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf2nw\" (UniqueName: \"kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302408 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302428 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302443 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4fv\" (UniqueName: \"kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302460 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302479 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302500 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302524 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.302548 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.381647 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.383284 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.384479 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.392137 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.392273 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-9cppk" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.392442 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.392463 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.403354 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-l6mcn"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404625 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404686 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404727 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404757 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404783 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf2nw\" (UniqueName: \"kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404857 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4fv\" (UniqueName: \"kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404875 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404892 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404916 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.404943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.417641 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.417899 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.417915 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.417975 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6l4dl" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.421999 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.425510 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l6mcn"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.426429 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.428130 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.429680 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.430667 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.431246 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.433024 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.434586 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.444452 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.447661 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.458781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4fv\" (UniqueName: \"kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv\") pod \"dnsmasq-dns-5c5cc7c5ff-gspw2\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.469423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf2nw\" (UniqueName: \"kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw\") pod \"keystone-bootstrap-q58ds\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.481845 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-55g4b"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.483303 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.488542 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cx4s4" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.488798 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.489004 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.505339 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.505948 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.505998 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506022 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506126 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506147 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f24tx\" (UniqueName: \"kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506167 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506186 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506240 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zhsn\" (UniqueName: \"kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506263 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.506281 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.517126 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-55g4b"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.525776 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.527177 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.563216 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.581023 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.596173 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-p9mq7"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.597183 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607246 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607288 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607328 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607350 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607378 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607419 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pztjx\" (UniqueName: \"kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607446 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrr4\" (UniqueName: \"kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607492 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607514 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f24tx\" (UniqueName: \"kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607534 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607556 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607574 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607597 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607619 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607634 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607652 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zhsn\" (UniqueName: \"kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607676 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.612286 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.615618 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607347 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.617140 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.607380 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dmgs8" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.621297 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.626421 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.630930 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p9mq7"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.653816 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.658738 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.660193 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.663757 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f24tx\" (UniqueName: \"kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx\") pod \"horizon-6bb9d9cd79-jbbgw\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.687909 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zhsn\" (UniqueName: \"kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn\") pod \"cinder-db-sync-l6mcn\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.705769 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.718747 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.718847 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.718882 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.718917 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.718963 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8gqc\" (UniqueName: \"kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719731 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719813 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719855 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719893 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pztjx\" (UniqueName: \"kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719940 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrr4\" (UniqueName: \"kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.719959 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.792840 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.795962 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.806888 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-6kwp6" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.807211 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.811610 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.812205 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.812338 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.816429 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.818485 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.822115 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.822215 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8gqc\" (UniqueName: \"kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.822248 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.833459 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.844064 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.845902 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.853713 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pztjx\" (UniqueName: \"kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx\") pod \"horizon-f64664c65-vk8pt\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.853890 4678 generic.go:334] "Generic (PLEG): container finished" podID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerID="164846cd982fa5935ac9132f26e56706fb81a779b26392cf73a44afe6b752fd2" exitCode=0 Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.853918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" event={"ID":"777f277f-fdea-4b55-bcbf-d8f6b7419656","Type":"ContainerDied","Data":"164846cd982fa5935ac9132f26e56706fb81a779b26392cf73a44afe6b752fd2"} Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.862119 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.866489 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.875640 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8gqc\" (UniqueName: \"kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc\") pod \"barbican-db-sync-p9mq7\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.886729 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p9mq7" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.894646 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrr4\" (UniqueName: \"kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4\") pod \"neutron-db-sync-55g4b\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.894752 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.909016 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.930874 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.933304 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948015 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st6xq\" (UniqueName: \"kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948070 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948129 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948195 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948223 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948251 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948288 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.948793 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.949981 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6mcn" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.967200 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.974500 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jb889"] Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.975693 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jb889" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.986667 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.986894 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.987098 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d46hs" Oct 13 12:59:23 crc kubenswrapper[4678]: I1013 12:59:23.987986 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.014313 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.018618 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.021907 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.026973 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054116 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054165 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054201 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st6xq\" (UniqueName: \"kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054224 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054265 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054290 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054311 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054353 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054392 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054438 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054462 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054480 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dvn4\" (UniqueName: \"kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054529 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.054973 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.058264 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.058719 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jb889"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.059152 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.074296 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.074692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.086201 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st6xq\" (UniqueName: \"kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.094014 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.100787 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.104399 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.106749 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.108709 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.109588 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.110659 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.125342 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.127146 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.156923 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.156979 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157007 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72tmg\" (UniqueName: \"kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157037 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157214 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157243 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157286 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157347 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157379 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157418 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157615 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157634 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157652 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157671 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157693 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157715 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157738 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157831 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157922 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.157950 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfvm5\" (UniqueName: \"kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158017 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158084 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158121 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158202 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158799 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.158896 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.159188 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.159243 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dvn4\" (UniqueName: \"kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.159262 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xf7w\" (UniqueName: \"kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.159761 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.186018 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.189163 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dvn4\" (UniqueName: \"kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4\") pod \"dnsmasq-dns-8b5c85b87-lt7cz\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.220477 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.247034 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272275 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272409 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272453 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272495 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272518 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9smh4\" (UniqueName: \"kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272562 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0\") pod \"777f277f-fdea-4b55-bcbf-d8f6b7419656\" (UID: \"777f277f-fdea-4b55-bcbf-d8f6b7419656\") " Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272786 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272821 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xf7w\" (UniqueName: \"kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272865 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272891 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272914 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72tmg\" (UniqueName: \"kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272937 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272957 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272976 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.272993 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273009 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273032 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273142 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273179 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273196 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273214 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273258 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273287 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.273321 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfvm5\" (UniqueName: \"kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.287843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.296359 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.296860 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.296981 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.297072 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.299084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.305505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfvm5\" (UniqueName: \"kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.305583 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.306018 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.306082 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.320670 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.321204 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72tmg\" (UniqueName: \"kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.321736 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.322087 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.322681 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.324304 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xf7w\" (UniqueName: \"kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.324496 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4" (OuterVolumeSpecName: "kube-api-access-9smh4") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "kube-api-access-9smh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.356963 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.375436 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9smh4\" (UniqueName: \"kubernetes.io/projected/777f277f-fdea-4b55-bcbf-d8f6b7419656-kube-api-access-9smh4\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.397519 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.398977 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.402201 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.404026 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data\") pod \"placement-db-sync-jb889\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.428442 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.431960 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q58ds"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.492031 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config" (OuterVolumeSpecName: "config") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.495754 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.519181 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.538234 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.546694 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.582276 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.582296 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.582309 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.582317 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.617565 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "777f277f-fdea-4b55-bcbf-d8f6b7419656" (UID: "777f277f-fdea-4b55-bcbf-d8f6b7419656"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.653331 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-p9mq7"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.684435 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/777f277f-fdea-4b55-bcbf-d8f6b7419656-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.687920 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jb889" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.698433 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.731272 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.779754 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.870269 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p9mq7" event={"ID":"04e988f0-2dc8-47d9-a8e0-65550e610593","Type":"ContainerStarted","Data":"61149d714b6758fbc242ae72f16f90eb3acd8d90229c9c089e604188d36c17ec"} Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.873752 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q58ds" event={"ID":"e8c111dd-b567-4700-89f3-164cc0b68a6d","Type":"ContainerStarted","Data":"41802d6786cb4f8cb51bcf3036ac42e4a2fbcee1d0a0e835568fe365fd496f29"} Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.875085 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" event={"ID":"3d97f26e-d62f-4287-82cd-3f44552fdd88","Type":"ContainerStarted","Data":"ea2883d0a685ba16566ba12d9d8e1365fc3d85f5a100475fc68dfb6acf23f4e7"} Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.876076 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerStarted","Data":"9e4775e82f993f29874255d7458f16590fa0324fd7e6675522ceb7c31b541a74"} Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.880682 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" event={"ID":"777f277f-fdea-4b55-bcbf-d8f6b7419656","Type":"ContainerDied","Data":"3384ae57541ed7f3a9e76477383d274b4a5227b8da04ec279b582e10c6119616"} Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.880712 4678 scope.go:117] "RemoveContainer" containerID="164846cd982fa5935ac9132f26e56706fb81a779b26392cf73a44afe6b752fd2" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.880839 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-zm25r" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.921097 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.936245 4678 scope.go:117] "RemoveContainer" containerID="e618c544475fd80f774a4aa8a1539945609241fc1c8cf2b3b1abc218d3522a2f" Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.938425 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:24 crc kubenswrapper[4678]: I1013 12:59:24.953899 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-zm25r"] Oct 13 12:59:24 crc kubenswrapper[4678]: W1013 12:59:24.993464 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d9788bd_6bed_4e7d_89c3_950a068a2c6d.slice/crio-f36212f0bb412201b6722469b788aa6e63f07f8c301dd662398246671ecd1491 WatchSource:0}: Error finding container f36212f0bb412201b6722469b788aa6e63f07f8c301dd662398246671ecd1491: Status 404 returned error can't find the container with id f36212f0bb412201b6722469b788aa6e63f07f8c301dd662398246671ecd1491 Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.235877 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-55g4b"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.247387 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l6mcn"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.301969 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.372388 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.406521 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: W1013 12:59:25.431065 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59f4d1c3_4733_46a9_8502_d03770bc285d.slice/crio-b5373016d9bd2599884ebd9ad166edc3491a32a475f99e283960b1f0669caac2 WatchSource:0}: Error finding container b5373016d9bd2599884ebd9ad166edc3491a32a475f99e283960b1f0669caac2: Status 404 returned error can't find the container with id b5373016d9bd2599884ebd9ad166edc3491a32a475f99e283960b1f0669caac2 Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.473952 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jb889"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.496784 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.636932 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.679648 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.722143 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 12:59:25 crc kubenswrapper[4678]: E1013 12:59:25.722971 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="init" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.722987 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="init" Oct 13 12:59:25 crc kubenswrapper[4678]: E1013 12:59:25.723069 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="dnsmasq-dns" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.723075 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="dnsmasq-dns" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.723326 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" containerName="dnsmasq-dns" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.724657 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.748096 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.757135 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.764764 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.834066 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.834137 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.834184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.834206 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb2nj\" (UniqueName: \"kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.834259 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.920990 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jb889" event={"ID":"5b7cde62-fcc9-40d6-8099-58838aea1ea7","Type":"ContainerStarted","Data":"1368bb93fa264defd2d1e4b2284c2316e9d519ba3522c0dabc02b1c0e188be8c"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.922780 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6mcn" event={"ID":"909be1eb-5445-4411-a2d0-c9e2618dd6bc","Type":"ContainerStarted","Data":"d074a156093305796441423107d27041fb59f228b237253c80641eba181ea36d"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.925116 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerStarted","Data":"ea7c993ac6465900c131ffa5a7890c782dd970950847e4f136f81492dcd032b8"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.928730 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" event={"ID":"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb","Type":"ContainerStarted","Data":"2a64971aeb4d86fdff35c990a089d65f8dd8ad5322ad9f69387b798ae197c98a"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.930854 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerStarted","Data":"f36212f0bb412201b6722469b788aa6e63f07f8c301dd662398246671ecd1491"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.935842 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.935928 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.936307 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.936340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb2nj\" (UniqueName: \"kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.937323 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.937627 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.938173 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.937176 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.942870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.948801 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-55g4b" event={"ID":"b72b73c0-6cc4-48fe-a32f-784786617f79","Type":"ContainerStarted","Data":"d41df3736e00c5b005c542f164030615478d676dcb300a5311cedea908a8c051"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.948855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-55g4b" event={"ID":"b72b73c0-6cc4-48fe-a32f-784786617f79","Type":"ContainerStarted","Data":"2aa98204c331c224af979d4e687e06d2b0e8ebee789bdf0c78380bd43fde4176"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.958458 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb2nj\" (UniqueName: \"kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj\") pod \"horizon-5bfb45cccf-j6vzn\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.961559 4678 generic.go:334] "Generic (PLEG): container finished" podID="3d97f26e-d62f-4287-82cd-3f44552fdd88" containerID="a05536cb19d35836e63a8381175e22c8b7f05ad0742ddf78ec12df8661eff3a8" exitCode=0 Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.961737 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" event={"ID":"3d97f26e-d62f-4287-82cd-3f44552fdd88","Type":"ContainerDied","Data":"a05536cb19d35836e63a8381175e22c8b7f05ad0742ddf78ec12df8661eff3a8"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.972236 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-55g4b" podStartSLOduration=2.972219166 podStartE2EDuration="2.972219166s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:25.967603095 +0000 UTC m=+954.052140979" watchObservedRunningTime="2025-10-13 12:59:25.972219166 +0000 UTC m=+954.056757050" Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.974444 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q58ds" event={"ID":"e8c111dd-b567-4700-89f3-164cc0b68a6d","Type":"ContainerStarted","Data":"ed14b44f1295610b341995a2d0a87e6a0584d0632b5b219cf3996b516f921889"} Oct 13 12:59:25 crc kubenswrapper[4678]: I1013 12:59:25.980171 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerStarted","Data":"eabcb46b0f5bd767833e924d79fccc38d8a3e758406528f37216ff67353d5984"} Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.015147 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerStarted","Data":"b5373016d9bd2599884ebd9ad166edc3491a32a475f99e283960b1f0669caac2"} Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.015752 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-q58ds" podStartSLOduration=3.015729935 podStartE2EDuration="3.015729935s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:26.015731685 +0000 UTC m=+954.100269559" watchObservedRunningTime="2025-10-13 12:59:26.015729935 +0000 UTC m=+954.100267829" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.103136 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.338773 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.461752 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn4fv\" (UniqueName: \"kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.461998 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.462134 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.462177 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.462260 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.462386 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb\") pod \"3d97f26e-d62f-4287-82cd-3f44552fdd88\" (UID: \"3d97f26e-d62f-4287-82cd-3f44552fdd88\") " Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.472291 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv" (OuterVolumeSpecName: "kube-api-access-qn4fv") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "kube-api-access-qn4fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.509032 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.509066 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.510311 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config" (OuterVolumeSpecName: "config") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.513703 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.517938 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d97f26e-d62f-4287-82cd-3f44552fdd88" (UID: "3d97f26e-d62f-4287-82cd-3f44552fdd88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.563961 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.564258 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn4fv\" (UniqueName: \"kubernetes.io/projected/3d97f26e-d62f-4287-82cd-3f44552fdd88-kube-api-access-qn4fv\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.564321 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.564371 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.564420 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.564469 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97f26e-d62f-4287-82cd-3f44552fdd88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.624509 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="777f277f-fdea-4b55-bcbf-d8f6b7419656" path="/var/lib/kubelet/pods/777f277f-fdea-4b55-bcbf-d8f6b7419656/volumes" Oct 13 12:59:26 crc kubenswrapper[4678]: I1013 12:59:26.631533 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.046658 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerStarted","Data":"33ba62f0f88f31acf8bf0e013a03c098daed5f8b3ce753f900e0bc6dac5535b2"} Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.049685 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerStarted","Data":"2518d676b50fc1c2732a42c26d1b9df5b4601ea707d7e1c2eb06a18ee4650ebb"} Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.051755 4678 generic.go:334] "Generic (PLEG): container finished" podID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerID="1434ac30094c34f35f70e8ac5b4750e89d2dcbcfdb8572489489dad7dfd3047e" exitCode=0 Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.051834 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" event={"ID":"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb","Type":"ContainerDied","Data":"1434ac30094c34f35f70e8ac5b4750e89d2dcbcfdb8572489489dad7dfd3047e"} Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.059617 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerStarted","Data":"0f8368914f3f4ee2fb0fd6444e6eeaf5be68327054789154ac03c48228a15c63"} Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.070209 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.070244 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-gspw2" event={"ID":"3d97f26e-d62f-4287-82cd-3f44552fdd88","Type":"ContainerDied","Data":"ea2883d0a685ba16566ba12d9d8e1365fc3d85f5a100475fc68dfb6acf23f4e7"} Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.070329 4678 scope.go:117] "RemoveContainer" containerID="a05536cb19d35836e63a8381175e22c8b7f05ad0742ddf78ec12df8661eff3a8" Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.117001 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:27 crc kubenswrapper[4678]: I1013 12:59:27.125499 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-gspw2"] Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.108389 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerStarted","Data":"da9ac05643775f177e92ba196cc5e36a8889b45de5fdaa6cd256b904fc1c1997"} Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.108920 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-log" containerID="cri-o://33ba62f0f88f31acf8bf0e013a03c098daed5f8b3ce753f900e0bc6dac5535b2" gracePeriod=30 Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.109558 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-httpd" containerID="cri-o://da9ac05643775f177e92ba196cc5e36a8889b45de5fdaa6cd256b904fc1c1997" gracePeriod=30 Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.114967 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" event={"ID":"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb","Type":"ContainerStarted","Data":"b2ba497c72150d0edafd0338454376d04aa4115cd2c884554cf877cc8debbbe0"} Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.115106 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.118980 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerStarted","Data":"dfd73890b80623aa90907c6d3ab182b3541e6786fdd780298301bd2720a93d6d"} Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.119116 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-httpd" containerID="cri-o://dfd73890b80623aa90907c6d3ab182b3541e6786fdd780298301bd2720a93d6d" gracePeriod=30 Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.119135 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-log" containerID="cri-o://0f8368914f3f4ee2fb0fd6444e6eeaf5be68327054789154ac03c48228a15c63" gracePeriod=30 Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.151085 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.151038687 podStartE2EDuration="5.151038687s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:28.125395916 +0000 UTC m=+956.209933800" watchObservedRunningTime="2025-10-13 12:59:28.151038687 +0000 UTC m=+956.235576571" Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.156926 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" podStartSLOduration=5.156907571 podStartE2EDuration="5.156907571s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:28.152235779 +0000 UTC m=+956.236773693" watchObservedRunningTime="2025-10-13 12:59:28.156907571 +0000 UTC m=+956.241445465" Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.186999 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.186981529 podStartE2EDuration="5.186981529s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:28.175819186 +0000 UTC m=+956.260357070" watchObservedRunningTime="2025-10-13 12:59:28.186981529 +0000 UTC m=+956.271519413" Oct 13 12:59:28 crc kubenswrapper[4678]: I1013 12:59:28.603875 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d97f26e-d62f-4287-82cd-3f44552fdd88" path="/var/lib/kubelet/pods/3d97f26e-d62f-4287-82cd-3f44552fdd88/volumes" Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.129624 4678 generic.go:334] "Generic (PLEG): container finished" podID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerID="dfd73890b80623aa90907c6d3ab182b3541e6786fdd780298301bd2720a93d6d" exitCode=0 Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.130018 4678 generic.go:334] "Generic (PLEG): container finished" podID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerID="0f8368914f3f4ee2fb0fd6444e6eeaf5be68327054789154ac03c48228a15c63" exitCode=143 Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.129706 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerDied","Data":"dfd73890b80623aa90907c6d3ab182b3541e6786fdd780298301bd2720a93d6d"} Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.130099 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerDied","Data":"0f8368914f3f4ee2fb0fd6444e6eeaf5be68327054789154ac03c48228a15c63"} Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.135427 4678 generic.go:334] "Generic (PLEG): container finished" podID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerID="da9ac05643775f177e92ba196cc5e36a8889b45de5fdaa6cd256b904fc1c1997" exitCode=0 Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.135465 4678 generic.go:334] "Generic (PLEG): container finished" podID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerID="33ba62f0f88f31acf8bf0e013a03c098daed5f8b3ce753f900e0bc6dac5535b2" exitCode=143 Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.137087 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerDied","Data":"da9ac05643775f177e92ba196cc5e36a8889b45de5fdaa6cd256b904fc1c1997"} Oct 13 12:59:29 crc kubenswrapper[4678]: I1013 12:59:29.137131 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerDied","Data":"33ba62f0f88f31acf8bf0e013a03c098daed5f8b3ce753f900e0bc6dac5535b2"} Oct 13 12:59:30 crc kubenswrapper[4678]: I1013 12:59:30.156627 4678 generic.go:334] "Generic (PLEG): container finished" podID="e8c111dd-b567-4700-89f3-164cc0b68a6d" containerID="ed14b44f1295610b341995a2d0a87e6a0584d0632b5b219cf3996b516f921889" exitCode=0 Oct 13 12:59:30 crc kubenswrapper[4678]: I1013 12:59:30.156804 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q58ds" event={"ID":"e8c111dd-b567-4700-89f3-164cc0b68a6d","Type":"ContainerDied","Data":"ed14b44f1295610b341995a2d0a87e6a0584d0632b5b219cf3996b516f921889"} Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.697710 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.774532 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 12:59:32 crc kubenswrapper[4678]: E1013 12:59:32.774971 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d97f26e-d62f-4287-82cd-3f44552fdd88" containerName="init" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.774989 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d97f26e-d62f-4287-82cd-3f44552fdd88" containerName="init" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.775183 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d97f26e-d62f-4287-82cd-3f44552fdd88" containerName="init" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.776179 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.793109 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.797755 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.810709 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.822345 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-94756fb8d-9h8cs"] Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.824140 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.856432 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-94756fb8d-9h8cs"] Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-combined-ca-bundle\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932281 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtwkn\" (UniqueName: \"kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932315 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-scripts\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932351 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932386 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932476 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932544 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932700 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-secret-key\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932873 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plj7c\" (UniqueName: \"kubernetes.io/projected/5bc24a68-de18-47d6-9573-67e978a05b05-kube-api-access-plj7c\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.932942 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bc24a68-de18-47d6-9573-67e978a05b05-logs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.933029 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-config-data\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.933139 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-tls-certs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.933289 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:32 crc kubenswrapper[4678]: I1013 12:59:32.933325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034554 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plj7c\" (UniqueName: \"kubernetes.io/projected/5bc24a68-de18-47d6-9573-67e978a05b05-kube-api-access-plj7c\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034598 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bc24a68-de18-47d6-9573-67e978a05b05-logs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034623 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-config-data\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034652 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-tls-certs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034686 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034702 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034748 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-combined-ca-bundle\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034776 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtwkn\" (UniqueName: \"kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034795 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-scripts\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034809 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034829 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034851 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034870 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.034892 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-secret-key\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.036127 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bc24a68-de18-47d6-9573-67e978a05b05-logs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.036872 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.037580 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.037594 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.037962 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-scripts\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.038514 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5bc24a68-de18-47d6-9573-67e978a05b05-config-data\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.045559 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-secret-key\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.045581 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-horizon-tls-certs\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.049659 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.050548 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.053125 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.053984 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plj7c\" (UniqueName: \"kubernetes.io/projected/5bc24a68-de18-47d6-9573-67e978a05b05-kube-api-access-plj7c\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.056364 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtwkn\" (UniqueName: \"kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn\") pod \"horizon-9bd58f858-cnpb4\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.067442 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bc24a68-de18-47d6-9573-67e978a05b05-combined-ca-bundle\") pod \"horizon-94756fb8d-9h8cs\" (UID: \"5bc24a68-de18-47d6-9573-67e978a05b05\") " pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.115191 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 12:59:33 crc kubenswrapper[4678]: I1013 12:59:33.161734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 12:59:34 crc kubenswrapper[4678]: I1013 12:59:34.325199 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 12:59:34 crc kubenswrapper[4678]: I1013 12:59:34.423078 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:59:34 crc kubenswrapper[4678]: I1013 12:59:34.423302 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" containerID="cri-o://2b39cfde21a792b1382e0e78dddc89f9877ffec87f645cbe7f81f7512b533f9e" gracePeriod=10 Oct 13 12:59:35 crc kubenswrapper[4678]: I1013 12:59:35.238433 4678 generic.go:334] "Generic (PLEG): container finished" podID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerID="2b39cfde21a792b1382e0e78dddc89f9877ffec87f645cbe7f81f7512b533f9e" exitCode=0 Oct 13 12:59:35 crc kubenswrapper[4678]: I1013 12:59:35.238815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" event={"ID":"bf2eadc6-a42c-4029-a644-3efd3b585d88","Type":"ContainerDied","Data":"2b39cfde21a792b1382e0e78dddc89f9877ffec87f645cbe7f81f7512b533f9e"} Oct 13 12:59:35 crc kubenswrapper[4678]: I1013 12:59:35.970347 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.096551 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf2nw\" (UniqueName: \"kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.096608 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.096686 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.096711 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.096761 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.097597 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle\") pod \"e8c111dd-b567-4700-89f3-164cc0b68a6d\" (UID: \"e8c111dd-b567-4700-89f3-164cc0b68a6d\") " Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.103265 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts" (OuterVolumeSpecName: "scripts") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.104419 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.104809 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.106744 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw" (OuterVolumeSpecName: "kube-api-access-rf2nw") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "kube-api-access-rf2nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.131572 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data" (OuterVolumeSpecName: "config-data") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.138066 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8c111dd-b567-4700-89f3-164cc0b68a6d" (UID: "e8c111dd-b567-4700-89f3-164cc0b68a6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200322 4678 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200362 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200375 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200387 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf2nw\" (UniqueName: \"kubernetes.io/projected/e8c111dd-b567-4700-89f3-164cc0b68a6d-kube-api-access-rf2nw\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200401 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.200412 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8c111dd-b567-4700-89f3-164cc0b68a6d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.246856 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q58ds" event={"ID":"e8c111dd-b567-4700-89f3-164cc0b68a6d","Type":"ContainerDied","Data":"41802d6786cb4f8cb51bcf3036ac42e4a2fbcee1d0a0e835568fe365fd496f29"} Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.246899 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41802d6786cb4f8cb51bcf3036ac42e4a2fbcee1d0a0e835568fe365fd496f29" Oct 13 12:59:36 crc kubenswrapper[4678]: I1013 12:59:36.246950 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q58ds" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.054861 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-q58ds"] Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.060733 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-q58ds"] Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.148233 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xldvz"] Oct 13 12:59:37 crc kubenswrapper[4678]: E1013 12:59:37.148644 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c111dd-b567-4700-89f3-164cc0b68a6d" containerName="keystone-bootstrap" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.148663 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c111dd-b567-4700-89f3-164cc0b68a6d" containerName="keystone-bootstrap" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.148843 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c111dd-b567-4700-89f3-164cc0b68a6d" containerName="keystone-bootstrap" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.149446 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.153200 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.153366 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.153484 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.153537 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xbmbf" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.155127 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xldvz"] Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.216951 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz6vj\" (UniqueName: \"kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.217029 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.217075 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.217092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.217122 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.217170 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.318726 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.318781 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.318830 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.318905 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.319030 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz6vj\" (UniqueName: \"kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.319119 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.322650 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.323433 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.323771 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.324660 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.331715 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.334646 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz6vj\" (UniqueName: \"kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj\") pod \"keystone-bootstrap-xldvz\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:37 crc kubenswrapper[4678]: I1013 12:59:37.498940 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xldvz" Oct 13 12:59:38 crc kubenswrapper[4678]: I1013 12:59:38.602971 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c111dd-b567-4700-89f3-164cc0b68a6d" path="/var/lib/kubelet/pods/e8c111dd-b567-4700-89f3-164cc0b68a6d/volumes" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.263256 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: connect: connection refused" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.628343 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766359 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766482 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766561 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766606 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72tmg\" (UniqueName: \"kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766633 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766693 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766745 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.766783 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs\") pod \"53b9b3ed-5668-4b48-957d-740ba25545e4\" (UID: \"53b9b3ed-5668-4b48-957d-740ba25545e4\") " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.767436 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.767485 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs" (OuterVolumeSpecName: "logs") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.772884 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts" (OuterVolumeSpecName: "scripts") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.773170 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg" (OuterVolumeSpecName: "kube-api-access-72tmg") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "kube-api-access-72tmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.783712 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.813757 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.817276 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data" (OuterVolumeSpecName: "config-data") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.835261 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "53b9b3ed-5668-4b48-957d-740ba25545e4" (UID: "53b9b3ed-5668-4b48-957d-740ba25545e4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868288 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868318 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72tmg\" (UniqueName: \"kubernetes.io/projected/53b9b3ed-5668-4b48-957d-740ba25545e4-kube-api-access-72tmg\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868331 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868339 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868348 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868356 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53b9b3ed-5668-4b48-957d-740ba25545e4-logs\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868385 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.868394 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53b9b3ed-5668-4b48-957d-740ba25545e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.884403 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 13 12:59:39 crc kubenswrapper[4678]: I1013 12:59:39.971034 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.277690 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53b9b3ed-5668-4b48-957d-740ba25545e4","Type":"ContainerDied","Data":"ea7c993ac6465900c131ffa5a7890c782dd970950847e4f136f81492dcd032b8"} Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.277743 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.277753 4678 scope.go:117] "RemoveContainer" containerID="da9ac05643775f177e92ba196cc5e36a8889b45de5fdaa6cd256b904fc1c1997" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.316266 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.335666 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.357272 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:40 crc kubenswrapper[4678]: E1013 12:59:40.357770 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-httpd" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.357974 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-httpd" Oct 13 12:59:40 crc kubenswrapper[4678]: E1013 12:59:40.358011 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-log" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.358020 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-log" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.358283 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-httpd" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.358310 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" containerName="glance-log" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.359524 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.362020 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.362570 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.365260 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.485862 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.485919 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.485961 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.485988 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.486021 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjltj\" (UniqueName: \"kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.486087 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.486125 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.486157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.587892 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.587944 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.587978 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.588230 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.588252 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.588290 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.588316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.588345 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjltj\" (UniqueName: \"kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.589555 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.590446 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.590573 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.606489 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjltj\" (UniqueName: \"kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.608538 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.615428 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.615597 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53b9b3ed-5668-4b48-957d-740ba25545e4" path="/var/lib/kubelet/pods/53b9b3ed-5668-4b48-957d-740ba25545e4/volumes" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.615885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.620103 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.635597 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " pod="openstack/glance-default-internal-api-0" Oct 13 12:59:40 crc kubenswrapper[4678]: I1013 12:59:40.676206 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.290697 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5fc4c96b-a1c5-464a-a301-8c14598428e6","Type":"ContainerDied","Data":"eabcb46b0f5bd767833e924d79fccc38d8a3e758406528f37216ff67353d5984"} Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.290752 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eabcb46b0f5bd767833e924d79fccc38d8a3e758406528f37216ff67353d5984" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.334858 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403429 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403482 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st6xq\" (UniqueName: \"kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403601 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403668 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403698 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403723 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.403801 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle\") pod \"5fc4c96b-a1c5-464a-a301-8c14598428e6\" (UID: \"5fc4c96b-a1c5-464a-a301-8c14598428e6\") " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.404326 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs" (OuterVolumeSpecName: "logs") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.404600 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.421805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.422184 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts" (OuterVolumeSpecName: "scripts") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.422669 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq" (OuterVolumeSpecName: "kube-api-access-st6xq") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "kube-api-access-st6xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.441690 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.457519 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.499388 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data" (OuterVolumeSpecName: "config-data") pod "5fc4c96b-a1c5-464a-a301-8c14598428e6" (UID: "5fc4c96b-a1c5-464a-a301-8c14598428e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505398 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505431 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505441 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-logs\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505451 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st6xq\" (UniqueName: \"kubernetes.io/projected/5fc4c96b-a1c5-464a-a301-8c14598428e6-kube-api-access-st6xq\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505464 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505491 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505502 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fc4c96b-a1c5-464a-a301-8c14598428e6-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.505511 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fc4c96b-a1c5-464a-a301-8c14598428e6-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.522329 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 13 12:59:41 crc kubenswrapper[4678]: I1013 12:59:41.607949 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.301458 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.342489 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.350368 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.377358 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:42 crc kubenswrapper[4678]: E1013 12:59:42.377844 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-log" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.377895 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-log" Oct 13 12:59:42 crc kubenswrapper[4678]: E1013 12:59:42.377954 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-httpd" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.377969 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-httpd" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.378305 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-httpd" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.378354 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" containerName="glance-log" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.382772 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.389884 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.390023 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.390021 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526560 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526686 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526714 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526757 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526799 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.526823 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b94c\" (UniqueName: \"kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.527006 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.527173 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.606263 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc4c96b-a1c5-464a-a301-8c14598428e6" path="/var/lib/kubelet/pods/5fc4c96b-a1c5-464a-a301-8c14598428e6/volumes" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629198 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629285 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629310 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b94c\" (UniqueName: \"kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629384 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629488 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.629716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.630407 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.630999 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.635310 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.636179 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.641740 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.642681 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.658331 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b94c\" (UniqueName: \"kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.659655 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " pod="openstack/glance-default-external-api-0" Oct 13 12:59:42 crc kubenswrapper[4678]: I1013 12:59:42.718949 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 12:59:45 crc kubenswrapper[4678]: I1013 12:59:45.333848 4678 generic.go:334] "Generic (PLEG): container finished" podID="b72b73c0-6cc4-48fe-a32f-784786617f79" containerID="d41df3736e00c5b005c542f164030615478d676dcb300a5311cedea908a8c051" exitCode=0 Oct 13 12:59:45 crc kubenswrapper[4678]: I1013 12:59:45.333931 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-55g4b" event={"ID":"b72b73c0-6cc4-48fe-a32f-784786617f79","Type":"ContainerDied","Data":"d41df3736e00c5b005c542f164030615478d676dcb300a5311cedea908a8c051"} Oct 13 12:59:49 crc kubenswrapper[4678]: I1013 12:59:49.262988 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Oct 13 12:59:49 crc kubenswrapper[4678]: E1013 12:59:49.782853 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 13 12:59:49 crc kubenswrapper[4678]: E1013 12:59:49.783358 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n8gqc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-p9mq7_openstack(04e988f0-2dc8-47d9-a8e0-65550e610593): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:59:49 crc kubenswrapper[4678]: E1013 12:59:49.784630 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-p9mq7" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" Oct 13 12:59:50 crc kubenswrapper[4678]: E1013 12:59:50.086482 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 13 12:59:50 crc kubenswrapper[4678]: E1013 12:59:50.086627 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n555h5dbh54ch588hfdh686h5fbh56fh568h5d9hb5hc9h649h56ch595h598h585h56h5ch559h67bh57ch5d6h544h699h59ch5c4h649h67h548h6dh54dq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xfvm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(59f4d1c3-4733-46a9-8502-d03770bc285d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.094960 4678 scope.go:117] "RemoveContainer" containerID="33ba62f0f88f31acf8bf0e013a03c098daed5f8b3ce753f900e0bc6dac5535b2" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.188513 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.204323 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.291900 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle\") pod \"b72b73c0-6cc4-48fe-a32f-784786617f79\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292000 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config\") pod \"b72b73c0-6cc4-48fe-a32f-784786617f79\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292065 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292128 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2l7x\" (UniqueName: \"kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292169 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292199 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292217 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292271 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrr4\" (UniqueName: \"kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4\") pod \"b72b73c0-6cc4-48fe-a32f-784786617f79\" (UID: \"b72b73c0-6cc4-48fe-a32f-784786617f79\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.292294 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config\") pod \"bf2eadc6-a42c-4029-a644-3efd3b585d88\" (UID: \"bf2eadc6-a42c-4029-a644-3efd3b585d88\") " Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.297520 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x" (OuterVolumeSpecName: "kube-api-access-h2l7x") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "kube-api-access-h2l7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.299441 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4" (OuterVolumeSpecName: "kube-api-access-hjrr4") pod "b72b73c0-6cc4-48fe-a32f-784786617f79" (UID: "b72b73c0-6cc4-48fe-a32f-784786617f79"). InnerVolumeSpecName "kube-api-access-hjrr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.319071 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b72b73c0-6cc4-48fe-a32f-784786617f79" (UID: "b72b73c0-6cc4-48fe-a32f-784786617f79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.321207 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config" (OuterVolumeSpecName: "config") pod "b72b73c0-6cc4-48fe-a32f-784786617f79" (UID: "b72b73c0-6cc4-48fe-a32f-784786617f79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.335429 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.339327 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.340175 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config" (OuterVolumeSpecName: "config") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.348410 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.359383 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf2eadc6-a42c-4029-a644-3efd3b585d88" (UID: "bf2eadc6-a42c-4029-a644-3efd3b585d88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.380340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" event={"ID":"bf2eadc6-a42c-4029-a644-3efd3b585d88","Type":"ContainerDied","Data":"3dc09dd2f31cf883cb75497ad60f402c5621b5d0a05d78cb8344bbbc7601c4ad"} Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.380418 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.392948 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-55g4b" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.393177 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-55g4b" event={"ID":"b72b73c0-6cc4-48fe-a32f-784786617f79","Type":"ContainerDied","Data":"2aa98204c331c224af979d4e687e06d2b0e8ebee789bdf0c78380bd43fde4176"} Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.393214 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aa98204c331c224af979d4e687e06d2b0e8ebee789bdf0c78380bd43fde4176" Oct 13 12:59:50 crc kubenswrapper[4678]: E1013 12:59:50.394009 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-p9mq7" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397024 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrr4\" (UniqueName: \"kubernetes.io/projected/b72b73c0-6cc4-48fe-a32f-784786617f79-kube-api-access-hjrr4\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397230 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397315 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397462 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b72b73c0-6cc4-48fe-a32f-784786617f79-config\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397543 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.397635 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2l7x\" (UniqueName: \"kubernetes.io/projected/bf2eadc6-a42c-4029-a644-3efd3b585d88-kube-api-access-h2l7x\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.399015 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.399208 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.399298 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf2eadc6-a42c-4029-a644-3efd3b585d88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.445298 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.454469 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-sg89w"] Oct 13 12:59:50 crc kubenswrapper[4678]: I1013 12:59:50.608712 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" path="/var/lib/kubelet/pods/bf2eadc6-a42c-4029-a644-3efd3b585d88/volumes" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.338178 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.338388 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zhsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-l6mcn_openstack(909be1eb-5445-4411-a2d0-c9e2618dd6bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.339864 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-l6mcn" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.368324 4678 scope.go:117] "RemoveContainer" containerID="2b39cfde21a792b1382e0e78dddc89f9877ffec87f645cbe7f81f7512b533f9e" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.419432 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-l6mcn" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500318 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.500729 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b72b73c0-6cc4-48fe-a32f-784786617f79" containerName="neutron-db-sync" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500748 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b72b73c0-6cc4-48fe-a32f-784786617f79" containerName="neutron-db-sync" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.500771 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500778 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" Oct 13 12:59:51 crc kubenswrapper[4678]: E1013 12:59:51.500791 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="init" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500797 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="init" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500954 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b72b73c0-6cc4-48fe-a32f-784786617f79" containerName="neutron-db-sync" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.500963 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.501888 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.527350 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.624479 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.624765 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.624798 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.625192 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmrzp\" (UniqueName: \"kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.625278 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.625461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.693671 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.697616 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.699733 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.699802 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.699992 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.701891 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cx4s4" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.711164 4678 scope.go:117] "RemoveContainer" containerID="b0762c9f3e8d151cff542a8ee34ae9ac467c1dccb24e346005a0ab1727940001" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.714760 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.726967 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.727025 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.727185 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmrzp\" (UniqueName: \"kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.727225 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.727296 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.727348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.728423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.729066 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.729534 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.729822 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.730013 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.747691 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmrzp\" (UniqueName: \"kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp\") pod \"dnsmasq-dns-84b966f6c9-k9xl9\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.831990 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.832035 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.832072 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25dwj\" (UniqueName: \"kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.832106 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.832140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.864143 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.933970 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.934262 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.934371 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.934395 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.934420 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25dwj\" (UniqueName: \"kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.938362 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.938705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.939314 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.941175 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.969168 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25dwj\" (UniqueName: \"kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj\") pod \"neutron-5c7b867ddd-k78g5\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:51 crc kubenswrapper[4678]: I1013 12:59:51.984315 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 12:59:51 crc kubenswrapper[4678]: W1013 12:59:51.985156 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4901114_e36f_4b5b_8d84_d64f02997156.slice/crio-c3a911472cc5c980932dc6d9528f853b779f07354bdeb1346939eb29f6c28736 WatchSource:0}: Error finding container c3a911472cc5c980932dc6d9528f853b779f07354bdeb1346939eb29f6c28736: Status 404 returned error can't find the container with id c3a911472cc5c980932dc6d9528f853b779f07354bdeb1346939eb29f6c28736 Oct 13 12:59:52 crc kubenswrapper[4678]: W1013 12:59:51.999587 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20c27b4c_8e65_490b_8ef7_aba250924b92.slice/crio-610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe WatchSource:0}: Error finding container 610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe: Status 404 returned error can't find the container with id 610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.001565 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xldvz"] Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.010494 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-94756fb8d-9h8cs"] Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.022875 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.284839 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 12:59:52 crc kubenswrapper[4678]: W1013 12:59:52.304576 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ce37dd_9cda_4d4f_9c9e_9c06dff80213.slice/crio-fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690 WatchSource:0}: Error finding container fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690: Status 404 returned error can't find the container with id fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.384414 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.440988 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.496477 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerStarted","Data":"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.496911 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerStarted","Data":"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.496768 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f64664c65-vk8pt" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon-log" containerID="cri-o://5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.497037 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f64664c65-vk8pt" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon" containerID="cri-o://236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.501929 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerStarted","Data":"c3a911472cc5c980932dc6d9528f853b779f07354bdeb1346939eb29f6c28736"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.505377 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xldvz" event={"ID":"20c27b4c-8e65-490b-8ef7-aba250924b92","Type":"ContainerStarted","Data":"bafc36f737e2055718d77357bfccba9a4187b0378c6d7b66cdd3a9b8fdcec96c"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.505455 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xldvz" event={"ID":"20c27b4c-8e65-490b-8ef7-aba250924b92","Type":"ContainerStarted","Data":"610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.523317 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f64664c65-vk8pt" podStartSLOduration=4.418884153 podStartE2EDuration="29.52330181s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="2025-10-13 12:59:25.011817955 +0000 UTC m=+953.096355829" lastFinishedPulling="2025-10-13 12:59:50.116235592 +0000 UTC m=+978.200773486" observedRunningTime="2025-10-13 12:59:52.517768715 +0000 UTC m=+980.602306599" watchObservedRunningTime="2025-10-13 12:59:52.52330181 +0000 UTC m=+980.607839694" Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.533582 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jb889" event={"ID":"5b7cde62-fcc9-40d6-8099-58838aea1ea7","Type":"ContainerStarted","Data":"083dc5b008aea5e16e3698c661030e91b999e9621bb64def235fa07d4fd8ed8b"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.538515 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerStarted","Data":"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.538695 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerStarted","Data":"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.538882 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bfb45cccf-j6vzn" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon-log" containerID="cri-o://a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.538970 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5bfb45cccf-j6vzn" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon" containerID="cri-o://a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.545952 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.546886 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerStarted","Data":"fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.568344 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xldvz" podStartSLOduration=15.568294538 podStartE2EDuration="15.568294538s" podCreationTimestamp="2025-10-13 12:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:52.534818731 +0000 UTC m=+980.619356615" watchObservedRunningTime="2025-10-13 12:59:52.568294538 +0000 UTC m=+980.652832432" Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.591755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerStarted","Data":"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.591823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerStarted","Data":"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.592022 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bb9d9cd79-jbbgw" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon-log" containerID="cri-o://dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.592188 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6bb9d9cd79-jbbgw" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon" containerID="cri-o://4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" gracePeriod=30 Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.592587 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jb889" podStartSLOduration=5.045559315 podStartE2EDuration="29.592562104s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="2025-10-13 12:59:25.563063731 +0000 UTC m=+953.647601615" lastFinishedPulling="2025-10-13 12:59:50.11006652 +0000 UTC m=+978.194604404" observedRunningTime="2025-10-13 12:59:52.561794928 +0000 UTC m=+980.646332812" watchObservedRunningTime="2025-10-13 12:59:52.592562104 +0000 UTC m=+980.677099988" Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.607778 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94756fb8d-9h8cs" event={"ID":"5bc24a68-de18-47d6-9573-67e978a05b05","Type":"ContainerStarted","Data":"d7715340216909343d8bd51b88071508b3d9e6b5c46a5a119d15cd3bc559d2f1"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.607824 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94756fb8d-9h8cs" event={"ID":"5bc24a68-de18-47d6-9573-67e978a05b05","Type":"ContainerStarted","Data":"d164f65764f9d68bc9512a2ab518b9159c3429129f30220be2a811ed42a19188"} Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.623014 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bfb45cccf-j6vzn" podStartSLOduration=2.960082537 podStartE2EDuration="27.622994471s" podCreationTimestamp="2025-10-13 12:59:25 +0000 UTC" firstStartedPulling="2025-10-13 12:59:26.680175457 +0000 UTC m=+954.764713351" lastFinishedPulling="2025-10-13 12:59:51.343087401 +0000 UTC m=+979.427625285" observedRunningTime="2025-10-13 12:59:52.612167417 +0000 UTC m=+980.696705311" watchObservedRunningTime="2025-10-13 12:59:52.622994471 +0000 UTC m=+980.707532355" Oct 13 12:59:52 crc kubenswrapper[4678]: I1013 12:59:52.739986 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6bb9d9cd79-jbbgw" podStartSLOduration=3.211948155 podStartE2EDuration="29.739969124s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="2025-10-13 12:59:24.792184773 +0000 UTC m=+952.876722657" lastFinishedPulling="2025-10-13 12:59:51.320205742 +0000 UTC m=+979.404743626" observedRunningTime="2025-10-13 12:59:52.734553042 +0000 UTC m=+980.819090926" watchObservedRunningTime="2025-10-13 12:59:52.739969124 +0000 UTC m=+980.824507008" Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.641836 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-94756fb8d-9h8cs" event={"ID":"5bc24a68-de18-47d6-9573-67e978a05b05","Type":"ContainerStarted","Data":"d6adf58f40ec8cf439e50d1c3c587ed8c13685c662721aa5fa3521731ed7f05a"} Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.654287 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerStarted","Data":"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71"} Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.660281 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" event={"ID":"47e524d5-b1ac-4559-af3a-b56fe79105a3","Type":"ContainerStarted","Data":"decb5a13bd5b1005b9bcd4b46913609f990798b73577cae84278638fce2399b4"} Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.666120 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerStarted","Data":"81ef72ca8bdc0f7c4c8c36e1b85ba85c4d0ca47b4defac9c5f057baff4fabc7c"} Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.668439 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerStarted","Data":"106e8ae00b24a86d3e2ed6c37b368a0e355bb5af8b1b28d82d9ec03fab07c0f1"} Oct 13 12:59:53 crc kubenswrapper[4678]: I1013 12:59:53.909999 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.125780 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f64664c65-vk8pt" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.264529 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-sg89w" podUID="bf2eadc6-a42c-4029-a644-3efd3b585d88" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.130:5353: i/o timeout" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.304815 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dcbfc8897-x7vwx"] Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.309585 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.315557 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dcbfc8897-x7vwx"] Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.352545 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.352559 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.405908 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-combined-ca-bundle\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.405953 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btgcd\" (UniqueName: \"kubernetes.io/projected/cc052259-a50b-4d53-ba5a-bbc8458aaa97-kube-api-access-btgcd\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.406019 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-ovndb-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.406060 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.406099 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-public-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.406139 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-internal-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.406156 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-httpd-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509219 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-combined-ca-bundle\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509262 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btgcd\" (UniqueName: \"kubernetes.io/projected/cc052259-a50b-4d53-ba5a-bbc8458aaa97-kube-api-access-btgcd\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509324 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-ovndb-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509351 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509392 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-public-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509429 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-internal-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.509445 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-httpd-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.515618 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-combined-ca-bundle\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.516866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-internal-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.519951 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.519975 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-public-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.520437 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-httpd-config\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.520831 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc052259-a50b-4d53-ba5a-bbc8458aaa97-ovndb-tls-certs\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.538965 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btgcd\" (UniqueName: \"kubernetes.io/projected/cc052259-a50b-4d53-ba5a-bbc8458aaa97-kube-api-access-btgcd\") pod \"neutron-dcbfc8897-x7vwx\" (UID: \"cc052259-a50b-4d53-ba5a-bbc8458aaa97\") " pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.676968 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.681002 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerStarted","Data":"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.685446 4678 generic.go:334] "Generic (PLEG): container finished" podID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerID="a96e400fd8cdafdb684c1a8687da680a24818d6b91fa0be598c91762dfcc4ae4" exitCode=0 Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.685571 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" event={"ID":"47e524d5-b1ac-4559-af3a-b56fe79105a3","Type":"ContainerDied","Data":"a96e400fd8cdafdb684c1a8687da680a24818d6b91fa0be598c91762dfcc4ae4"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.691148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerStarted","Data":"5a3622e06cf6ea08cad1728893e56ff53651623b66d8e1a4d9834187a0431427"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.692574 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerStarted","Data":"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.693669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerStarted","Data":"981a5d7b748bc3ac9dd71d235f47c8d951188cc2a02481a5101f52897a4effc8"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.693784 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerStarted","Data":"c72e762a63fc27cfa9694add9b2859136f37527127fc566769862abbb1e9dd6b"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.694341 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.700875 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerStarted","Data":"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e"} Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.710585 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-9bd58f858-cnpb4" podStartSLOduration=22.710560581 podStartE2EDuration="22.710560581s" podCreationTimestamp="2025-10-13 12:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:54.701212006 +0000 UTC m=+982.785749890" watchObservedRunningTime="2025-10-13 12:59:54.710560581 +0000 UTC m=+982.795098465" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.726304 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-94756fb8d-9h8cs" podStartSLOduration=22.726284293 podStartE2EDuration="22.726284293s" podCreationTimestamp="2025-10-13 12:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:54.721451627 +0000 UTC m=+982.805989521" watchObservedRunningTime="2025-10-13 12:59:54.726284293 +0000 UTC m=+982.810822177" Oct 13 12:59:54 crc kubenswrapper[4678]: I1013 12:59:54.742257 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c7b867ddd-k78g5" podStartSLOduration=3.742233781 podStartE2EDuration="3.742233781s" podCreationTimestamp="2025-10-13 12:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:54.735951586 +0000 UTC m=+982.820489470" watchObservedRunningTime="2025-10-13 12:59:54.742233781 +0000 UTC m=+982.826771665" Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.104340 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.212840 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dcbfc8897-x7vwx"] Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.732547 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerStarted","Data":"51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.733970 4678 generic.go:334] "Generic (PLEG): container finished" podID="5b7cde62-fcc9-40d6-8099-58838aea1ea7" containerID="083dc5b008aea5e16e3698c661030e91b999e9621bb64def235fa07d4fd8ed8b" exitCode=0 Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.734029 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jb889" event={"ID":"5b7cde62-fcc9-40d6-8099-58838aea1ea7","Type":"ContainerDied","Data":"083dc5b008aea5e16e3698c661030e91b999e9621bb64def235fa07d4fd8ed8b"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.736841 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerStarted","Data":"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.749556 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dcbfc8897-x7vwx" event={"ID":"cc052259-a50b-4d53-ba5a-bbc8458aaa97","Type":"ContainerStarted","Data":"f2125ed7126686f0b96ff828b586c1cfcce7f943a9a4427ea1a28bdabe7e7e56"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.749606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dcbfc8897-x7vwx" event={"ID":"cc052259-a50b-4d53-ba5a-bbc8458aaa97","Type":"ContainerStarted","Data":"33ed1c4a23e3372a83a5163c922b0593708b1dde013e09f63af9d60617378a0a"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.751853 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" event={"ID":"47e524d5-b1ac-4559-af3a-b56fe79105a3","Type":"ContainerStarted","Data":"0ed8f5ca616bc1e655ccf3d1bc269a84f3ea4b5e468b876edb9134acb7cb6fea"} Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.767147 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=14.767127881 podStartE2EDuration="14.767127881s" podCreationTimestamp="2025-10-13 12:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:56.756439631 +0000 UTC m=+984.840977515" watchObservedRunningTime="2025-10-13 12:59:56.767127881 +0000 UTC m=+984.851665765" Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.785927 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=16.785907763 podStartE2EDuration="16.785907763s" podCreationTimestamp="2025-10-13 12:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:56.778392336 +0000 UTC m=+984.862930220" watchObservedRunningTime="2025-10-13 12:59:56.785907763 +0000 UTC m=+984.870445637" Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.817626 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" podStartSLOduration=5.817602953 podStartE2EDuration="5.817602953s" podCreationTimestamp="2025-10-13 12:59:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:56.813437144 +0000 UTC m=+984.897975028" watchObservedRunningTime="2025-10-13 12:59:56.817602953 +0000 UTC m=+984.902140837" Oct 13 12:59:56 crc kubenswrapper[4678]: I1013 12:59:56.864545 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 12:59:57 crc kubenswrapper[4678]: I1013 12:59:57.761691 4678 generic.go:334] "Generic (PLEG): container finished" podID="20c27b4c-8e65-490b-8ef7-aba250924b92" containerID="bafc36f737e2055718d77357bfccba9a4187b0378c6d7b66cdd3a9b8fdcec96c" exitCode=0 Oct 13 12:59:57 crc kubenswrapper[4678]: I1013 12:59:57.761742 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xldvz" event={"ID":"20c27b4c-8e65-490b-8ef7-aba250924b92","Type":"ContainerDied","Data":"bafc36f737e2055718d77357bfccba9a4187b0378c6d7b66cdd3a9b8fdcec96c"} Oct 13 12:59:57 crc kubenswrapper[4678]: I1013 12:59:57.765583 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dcbfc8897-x7vwx" event={"ID":"cc052259-a50b-4d53-ba5a-bbc8458aaa97","Type":"ContainerStarted","Data":"27715d4270a1eb9039e6808761e45fb342ee6b272857138d9e4643f1c2ecffed"} Oct 13 12:59:57 crc kubenswrapper[4678]: I1013 12:59:57.767119 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 12:59:57 crc kubenswrapper[4678]: I1013 12:59:57.806126 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dcbfc8897-x7vwx" podStartSLOduration=3.806108402 podStartE2EDuration="3.806108402s" podCreationTimestamp="2025-10-13 12:59:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 12:59:57.796644794 +0000 UTC m=+985.881182678" watchObservedRunningTime="2025-10-13 12:59:57.806108402 +0000 UTC m=+985.890646286" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.116870 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jb889" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.179704 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle\") pod \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.179787 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data\") pod \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.179865 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs\") pod \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.179881 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts\") pod \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.179918 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xf7w\" (UniqueName: \"kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w\") pod \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\" (UID: \"5b7cde62-fcc9-40d6-8099-58838aea1ea7\") " Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.182531 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs" (OuterVolumeSpecName: "logs") pod "5b7cde62-fcc9-40d6-8099-58838aea1ea7" (UID: "5b7cde62-fcc9-40d6-8099-58838aea1ea7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.187798 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts" (OuterVolumeSpecName: "scripts") pod "5b7cde62-fcc9-40d6-8099-58838aea1ea7" (UID: "5b7cde62-fcc9-40d6-8099-58838aea1ea7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.200249 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w" (OuterVolumeSpecName: "kube-api-access-8xf7w") pod "5b7cde62-fcc9-40d6-8099-58838aea1ea7" (UID: "5b7cde62-fcc9-40d6-8099-58838aea1ea7"). InnerVolumeSpecName "kube-api-access-8xf7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.216366 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b7cde62-fcc9-40d6-8099-58838aea1ea7" (UID: "5b7cde62-fcc9-40d6-8099-58838aea1ea7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.289970 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.290013 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b7cde62-fcc9-40d6-8099-58838aea1ea7-logs\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.290022 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.290031 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xf7w\" (UniqueName: \"kubernetes.io/projected/5b7cde62-fcc9-40d6-8099-58838aea1ea7-kube-api-access-8xf7w\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.295185 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data" (OuterVolumeSpecName: "config-data") pod "5b7cde62-fcc9-40d6-8099-58838aea1ea7" (UID: "5b7cde62-fcc9-40d6-8099-58838aea1ea7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.391393 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b7cde62-fcc9-40d6-8099-58838aea1ea7-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.780242 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jb889" event={"ID":"5b7cde62-fcc9-40d6-8099-58838aea1ea7","Type":"ContainerDied","Data":"1368bb93fa264defd2d1e4b2284c2316e9d519ba3522c0dabc02b1c0e188be8c"} Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.781104 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1368bb93fa264defd2d1e4b2284c2316e9d519ba3522c0dabc02b1c0e188be8c" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.780306 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jb889" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.963702 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-694bd84568-p2blx"] Oct 13 12:59:58 crc kubenswrapper[4678]: E1013 12:59:58.964331 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b7cde62-fcc9-40d6-8099-58838aea1ea7" containerName="placement-db-sync" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.964348 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b7cde62-fcc9-40d6-8099-58838aea1ea7" containerName="placement-db-sync" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.964513 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b7cde62-fcc9-40d6-8099-58838aea1ea7" containerName="placement-db-sync" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.967116 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.975965 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.980041 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.980251 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.980360 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-d46hs" Oct 13 12:59:58 crc kubenswrapper[4678]: I1013 12:59:58.980453 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.021238 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-694bd84568-p2blx"] Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104268 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-public-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104331 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-combined-ca-bundle\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104353 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-internal-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104598 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b5cb1-0247-44a1-b63b-cffb9a30e588-logs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104660 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-config-data\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104766 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfwpw\" (UniqueName: \"kubernetes.io/projected/776b5cb1-0247-44a1-b63b-cffb9a30e588-kube-api-access-pfwpw\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.104960 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-scripts\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.207840 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-combined-ca-bundle\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.207908 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-internal-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.208011 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b5cb1-0247-44a1-b63b-cffb9a30e588-logs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.208145 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-config-data\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.208216 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfwpw\" (UniqueName: \"kubernetes.io/projected/776b5cb1-0247-44a1-b63b-cffb9a30e588-kube-api-access-pfwpw\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.208297 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-scripts\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.208583 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/776b5cb1-0247-44a1-b63b-cffb9a30e588-logs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.209215 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-public-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.212465 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-internal-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.214221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-public-tls-certs\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.214885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-combined-ca-bundle\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.214962 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-config-data\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.217703 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/776b5cb1-0247-44a1-b63b-cffb9a30e588-scripts\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.224604 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfwpw\" (UniqueName: \"kubernetes.io/projected/776b5cb1-0247-44a1-b63b-cffb9a30e588-kube-api-access-pfwpw\") pod \"placement-694bd84568-p2blx\" (UID: \"776b5cb1-0247-44a1-b63b-cffb9a30e588\") " pod="openstack/placement-694bd84568-p2blx" Oct 13 12:59:59 crc kubenswrapper[4678]: I1013 12:59:59.306846 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-694bd84568-p2blx" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.134741 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4"] Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.136440 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.145783 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4"] Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.149455 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.149598 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.228091 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.228208 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.228262 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88nk\" (UniqueName: \"kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.329078 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.329177 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.329230 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88nk\" (UniqueName: \"kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.330619 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.335511 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.348607 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88nk\" (UniqueName: \"kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk\") pod \"collect-profiles-29339340-mzxk4\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.465595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.676941 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.677247 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.723301 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.754655 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.812253 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xldvz" event={"ID":"20c27b4c-8e65-490b-8ef7-aba250924b92","Type":"ContainerDied","Data":"610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe"} Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.812287 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="610e4d8ad61715b0d5ace37d631f0b89de59235db30be7319085a81398b243fe" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.812320 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.812342 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.840491 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xldvz" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.943853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.943992 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.944180 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.944215 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.944239 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz6vj\" (UniqueName: \"kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.944263 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys\") pod \"20c27b4c-8e65-490b-8ef7-aba250924b92\" (UID: \"20c27b4c-8e65-490b-8ef7-aba250924b92\") " Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.950447 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts" (OuterVolumeSpecName: "scripts") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.951256 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.955141 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.989821 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj" (OuterVolumeSpecName: "kube-api-access-kz6vj") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "kube-api-access-kz6vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:00 crc kubenswrapper[4678]: I1013 13:00:00.990582 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.010659 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data" (OuterVolumeSpecName: "config-data") pod "20c27b4c-8e65-490b-8ef7-aba250924b92" (UID: "20c27b4c-8e65-490b-8ef7-aba250924b92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047239 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047266 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047277 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047286 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047296 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz6vj\" (UniqueName: \"kubernetes.io/projected/20c27b4c-8e65-490b-8ef7-aba250924b92-kube-api-access-kz6vj\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.047306 4678 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/20c27b4c-8e65-490b-8ef7-aba250924b92-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.819613 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xldvz" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.873276 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.921655 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 13:00:01 crc kubenswrapper[4678]: I1013 13:00:01.921866 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="dnsmasq-dns" containerID="cri-o://b2ba497c72150d0edafd0338454376d04aa4115cd2c884554cf877cc8debbbe0" gracePeriod=10 Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.010742 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b4bffbf65-fc8vg"] Oct 13 13:00:02 crc kubenswrapper[4678]: E1013 13:00:02.011576 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c27b4c-8e65-490b-8ef7-aba250924b92" containerName="keystone-bootstrap" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.011661 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c27b4c-8e65-490b-8ef7-aba250924b92" containerName="keystone-bootstrap" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.011920 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c27b4c-8e65-490b-8ef7-aba250924b92" containerName="keystone-bootstrap" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.012522 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.018096 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.018310 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.018411 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xbmbf" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.018517 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.018639 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.019318 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.051113 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b4bffbf65-fc8vg"] Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066649 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-fernet-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066715 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-internal-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066759 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-credential-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066815 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-combined-ca-bundle\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066833 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-public-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.066943 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-scripts\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.067361 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-config-data\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.067438 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk6ns\" (UniqueName: \"kubernetes.io/projected/f6470c70-b6b8-456a-9ca0-5bc9c9122674-kube-api-access-kk6ns\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169615 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk6ns\" (UniqueName: \"kubernetes.io/projected/f6470c70-b6b8-456a-9ca0-5bc9c9122674-kube-api-access-kk6ns\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169685 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-fernet-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-internal-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169732 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-credential-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169761 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-combined-ca-bundle\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169774 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-public-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169799 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-scripts\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.169850 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-config-data\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.175116 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-config-data\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.181528 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-fernet-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.182582 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-public-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.188955 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-internal-tls-certs\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.189520 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-credential-keys\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.194258 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-scripts\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.204373 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6470c70-b6b8-456a-9ca0-5bc9c9122674-combined-ca-bundle\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.224867 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk6ns\" (UniqueName: \"kubernetes.io/projected/f6470c70-b6b8-456a-9ca0-5bc9c9122674-kube-api-access-kk6ns\") pod \"keystone-b4bffbf65-fc8vg\" (UID: \"f6470c70-b6b8-456a-9ca0-5bc9c9122674\") " pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.354165 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.719117 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.719929 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.814874 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.848273 4678 generic.go:334] "Generic (PLEG): container finished" podID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerID="b2ba497c72150d0edafd0338454376d04aa4115cd2c884554cf877cc8debbbe0" exitCode=0 Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.848337 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" event={"ID":"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb","Type":"ContainerDied","Data":"b2ba497c72150d0edafd0338454376d04aa4115cd2c884554cf877cc8debbbe0"} Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.848433 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.848804 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 13:00:02 crc kubenswrapper[4678]: I1013 13:00:02.882791 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.116487 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.116546 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.119820 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.162383 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.162689 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.165833 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-94756fb8d-9h8cs" podUID="5bc24a68-de18-47d6-9573-67e978a05b05" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.411451 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.413005 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:03 crc kubenswrapper[4678]: I1013 13:00:03.859273 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.221269 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.328722 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.328759 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dvn4\" (UniqueName: \"kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.328787 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.328894 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.328926 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.329029 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config\") pod \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\" (UID: \"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb\") " Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.364291 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4" (OuterVolumeSpecName: "kube-api-access-9dvn4") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "kube-api-access-9dvn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.431740 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dvn4\" (UniqueName: \"kubernetes.io/projected/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-kube-api-access-9dvn4\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.522809 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-694bd84568-p2blx"] Oct 13 13:00:04 crc kubenswrapper[4678]: W1013 13:00:04.536269 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod776b5cb1_0247_44a1_b63b_cffb9a30e588.slice/crio-cd300d627d609fe78f0f623e9f6038bd7fb3b8ea8ba952141bc5915f2c4d5125 WatchSource:0}: Error finding container cd300d627d609fe78f0f623e9f6038bd7fb3b8ea8ba952141bc5915f2c4d5125: Status 404 returned error can't find the container with id cd300d627d609fe78f0f623e9f6038bd7fb3b8ea8ba952141bc5915f2c4d5125 Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.586167 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4"] Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.628667 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.635592 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.647568 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.657244 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.676570 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config" (OuterVolumeSpecName: "config") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.694630 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" (UID: "ee4d3b9f-8933-4aa5-a660-90cbddd6fccb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.703334 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b4bffbf65-fc8vg"] Oct 13 13:00:04 crc kubenswrapper[4678]: W1013 13:00:04.710886 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6470c70_b6b8_456a_9ca0_5bc9c9122674.slice/crio-0769c2a7b3f3f5a81b94aa8d4d3a9f35a470d2e2c8f70e14ab11866666418a21 WatchSource:0}: Error finding container 0769c2a7b3f3f5a81b94aa8d4d3a9f35a470d2e2c8f70e14ab11866666418a21: Status 404 returned error can't find the container with id 0769c2a7b3f3f5a81b94aa8d4d3a9f35a470d2e2c8f70e14ab11866666418a21 Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.737446 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.737480 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.737491 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.737502 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.886024 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-694bd84568-p2blx" event={"ID":"776b5cb1-0247-44a1-b63b-cffb9a30e588","Type":"ContainerStarted","Data":"693c19b7a8ea4f3b471014d734f22e8443299242172b1c32e69d112ed2f6d819"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.887423 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-694bd84568-p2blx" event={"ID":"776b5cb1-0247-44a1-b63b-cffb9a30e588","Type":"ContainerStarted","Data":"cd300d627d609fe78f0f623e9f6038bd7fb3b8ea8ba952141bc5915f2c4d5125"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.929140 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p9mq7" event={"ID":"04e988f0-2dc8-47d9-a8e0-65550e610593","Type":"ContainerStarted","Data":"b361fadc7b71f790f9e8923b00536175892297f49ef4481560a3a3fb3d55374b"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.941019 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" event={"ID":"e0124058-8b2a-4a81-80ba-34fd9b93fd2b","Type":"ContainerStarted","Data":"4be9a46d6d2f64e3a281cd836a6d0f18806303022ed9aaf9245aa9eb6a437499"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.950179 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" event={"ID":"ee4d3b9f-8933-4aa5-a660-90cbddd6fccb","Type":"ContainerDied","Data":"2a64971aeb4d86fdff35c990a089d65f8dd8ad5322ad9f69387b798ae197c98a"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.950227 4678 scope.go:117] "RemoveContainer" containerID="b2ba497c72150d0edafd0338454376d04aa4115cd2c884554cf877cc8debbbe0" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.950341 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-lt7cz" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.950815 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-p9mq7" podStartSLOduration=2.494920237 podStartE2EDuration="41.950792034s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="2025-10-13 12:59:24.687265765 +0000 UTC m=+952.771803649" lastFinishedPulling="2025-10-13 13:00:04.143137562 +0000 UTC m=+992.227675446" observedRunningTime="2025-10-13 13:00:04.947937009 +0000 UTC m=+993.032474893" watchObservedRunningTime="2025-10-13 13:00:04.950792034 +0000 UTC m=+993.035329918" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.979088 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" podStartSLOduration=4.979069094 podStartE2EDuration="4.979069094s" podCreationTimestamp="2025-10-13 13:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:04.965461798 +0000 UTC m=+993.049999682" watchObservedRunningTime="2025-10-13 13:00:04.979069094 +0000 UTC m=+993.063606978" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.980755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b4bffbf65-fc8vg" event={"ID":"f6470c70-b6b8-456a-9ca0-5bc9c9122674","Type":"ContainerStarted","Data":"0769c2a7b3f3f5a81b94aa8d4d3a9f35a470d2e2c8f70e14ab11866666418a21"} Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.987311 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:00:04 crc kubenswrapper[4678]: I1013 13:00:04.987939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerStarted","Data":"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed"} Oct 13 13:00:05 crc kubenswrapper[4678]: I1013 13:00:05.104076 4678 scope.go:117] "RemoveContainer" containerID="1434ac30094c34f35f70e8ac5b4750e89d2dcbcfdb8572489489dad7dfd3047e" Oct 13 13:00:05 crc kubenswrapper[4678]: I1013 13:00:05.199141 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 13:00:05 crc kubenswrapper[4678]: I1013 13:00:05.207279 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-lt7cz"] Oct 13 13:00:05 crc kubenswrapper[4678]: I1013 13:00:05.259803 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 13:00:05 crc kubenswrapper[4678]: I1013 13:00:05.998607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-694bd84568-p2blx" event={"ID":"776b5cb1-0247-44a1-b63b-cffb9a30e588","Type":"ContainerStarted","Data":"65e9d18618cab941eeed5377c2b95956bda5e7eb607a88104e985768eb39cb3c"} Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.000184 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-694bd84568-p2blx" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.000223 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-694bd84568-p2blx" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.002593 4678 generic.go:334] "Generic (PLEG): container finished" podID="e0124058-8b2a-4a81-80ba-34fd9b93fd2b" containerID="e9356078dd51f50c6c8c99a3c3b0c2a419e303f89dbffcbf3ac71ef52e1f353d" exitCode=0 Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.002643 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" event={"ID":"e0124058-8b2a-4a81-80ba-34fd9b93fd2b","Type":"ContainerDied","Data":"e9356078dd51f50c6c8c99a3c3b0c2a419e303f89dbffcbf3ac71ef52e1f353d"} Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.004855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6mcn" event={"ID":"909be1eb-5445-4411-a2d0-c9e2618dd6bc","Type":"ContainerStarted","Data":"dc3b37825d146cb808e7264a0b62e84dc6031e46bd732274f224c450da6148aa"} Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.009287 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.011129 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b4bffbf65-fc8vg" event={"ID":"f6470c70-b6b8-456a-9ca0-5bc9c9122674","Type":"ContainerStarted","Data":"da0b1833676322ab9e15bc22723fab30a0893c03f56993da4d0ac30ac8a04f53"} Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.011165 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.028439 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-694bd84568-p2blx" podStartSLOduration=8.028418245 podStartE2EDuration="8.028418245s" podCreationTimestamp="2025-10-13 12:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:06.026508055 +0000 UTC m=+994.111045949" watchObservedRunningTime="2025-10-13 13:00:06.028418245 +0000 UTC m=+994.112956129" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.045251 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-l6mcn" podStartSLOduration=3.617916666 podStartE2EDuration="43.045230805s" podCreationTimestamp="2025-10-13 12:59:23 +0000 UTC" firstStartedPulling="2025-10-13 12:59:25.249795217 +0000 UTC m=+953.334333091" lastFinishedPulling="2025-10-13 13:00:04.677109346 +0000 UTC m=+992.761647230" observedRunningTime="2025-10-13 13:00:06.039944407 +0000 UTC m=+994.124482291" watchObservedRunningTime="2025-10-13 13:00:06.045230805 +0000 UTC m=+994.129768699" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.079150 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-b4bffbf65-fc8vg" podStartSLOduration=5.079132993 podStartE2EDuration="5.079132993s" podCreationTimestamp="2025-10-13 13:00:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:06.069762398 +0000 UTC m=+994.154300282" watchObservedRunningTime="2025-10-13 13:00:06.079132993 +0000 UTC m=+994.163670877" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.513695 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 13:00:06 crc kubenswrapper[4678]: I1013 13:00:06.605623 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" path="/var/lib/kubelet/pods/ee4d3b9f-8933-4aa5-a660-90cbddd6fccb/volumes" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.363929 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.489698 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t88nk\" (UniqueName: \"kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk\") pod \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.489856 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume\") pod \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.489954 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume\") pod \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\" (UID: \"e0124058-8b2a-4a81-80ba-34fd9b93fd2b\") " Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.490564 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e0124058-8b2a-4a81-80ba-34fd9b93fd2b" (UID: "e0124058-8b2a-4a81-80ba-34fd9b93fd2b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.509244 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk" (OuterVolumeSpecName: "kube-api-access-t88nk") pod "e0124058-8b2a-4a81-80ba-34fd9b93fd2b" (UID: "e0124058-8b2a-4a81-80ba-34fd9b93fd2b"). InnerVolumeSpecName "kube-api-access-t88nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.511261 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e0124058-8b2a-4a81-80ba-34fd9b93fd2b" (UID: "e0124058-8b2a-4a81-80ba-34fd9b93fd2b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.592327 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t88nk\" (UniqueName: \"kubernetes.io/projected/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-kube-api-access-t88nk\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.592357 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:07 crc kubenswrapper[4678]: I1013 13:00:07.592367 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e0124058-8b2a-4a81-80ba-34fd9b93fd2b-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:08 crc kubenswrapper[4678]: I1013 13:00:08.029624 4678 generic.go:334] "Generic (PLEG): container finished" podID="04e988f0-2dc8-47d9-a8e0-65550e610593" containerID="b361fadc7b71f790f9e8923b00536175892297f49ef4481560a3a3fb3d55374b" exitCode=0 Oct 13 13:00:08 crc kubenswrapper[4678]: I1013 13:00:08.029667 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p9mq7" event={"ID":"04e988f0-2dc8-47d9-a8e0-65550e610593","Type":"ContainerDied","Data":"b361fadc7b71f790f9e8923b00536175892297f49ef4481560a3a3fb3d55374b"} Oct 13 13:00:08 crc kubenswrapper[4678]: I1013 13:00:08.031732 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" event={"ID":"e0124058-8b2a-4a81-80ba-34fd9b93fd2b","Type":"ContainerDied","Data":"4be9a46d6d2f64e3a281cd836a6d0f18806303022ed9aaf9245aa9eb6a437499"} Oct 13 13:00:08 crc kubenswrapper[4678]: I1013 13:00:08.031797 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4be9a46d6d2f64e3a281cd836a6d0f18806303022ed9aaf9245aa9eb6a437499" Oct 13 13:00:08 crc kubenswrapper[4678]: I1013 13:00:08.031763 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.455499 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p9mq7" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.527213 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8gqc\" (UniqueName: \"kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc\") pod \"04e988f0-2dc8-47d9-a8e0-65550e610593\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.527279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle\") pod \"04e988f0-2dc8-47d9-a8e0-65550e610593\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.527331 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data\") pod \"04e988f0-2dc8-47d9-a8e0-65550e610593\" (UID: \"04e988f0-2dc8-47d9-a8e0-65550e610593\") " Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.536150 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "04e988f0-2dc8-47d9-a8e0-65550e610593" (UID: "04e988f0-2dc8-47d9-a8e0-65550e610593"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.557295 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc" (OuterVolumeSpecName: "kube-api-access-n8gqc") pod "04e988f0-2dc8-47d9-a8e0-65550e610593" (UID: "04e988f0-2dc8-47d9-a8e0-65550e610593"). InnerVolumeSpecName "kube-api-access-n8gqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.562167 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04e988f0-2dc8-47d9-a8e0-65550e610593" (UID: "04e988f0-2dc8-47d9-a8e0-65550e610593"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.629311 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8gqc\" (UniqueName: \"kubernetes.io/projected/04e988f0-2dc8-47d9-a8e0-65550e610593-kube-api-access-n8gqc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.629339 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:09 crc kubenswrapper[4678]: I1013 13:00:09.629348 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04e988f0-2dc8-47d9-a8e0-65550e610593-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.052357 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-p9mq7" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.052351 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-p9mq7" event={"ID":"04e988f0-2dc8-47d9-a8e0-65550e610593","Type":"ContainerDied","Data":"61149d714b6758fbc242ae72f16f90eb3acd8d90229c9c089e604188d36c17ec"} Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.052465 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61149d714b6758fbc242ae72f16f90eb3acd8d90229c9c089e604188d36c17ec" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.078826 4678 generic.go:334] "Generic (PLEG): container finished" podID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" containerID="dc3b37825d146cb808e7264a0b62e84dc6031e46bd732274f224c450da6148aa" exitCode=0 Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.078869 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6mcn" event={"ID":"909be1eb-5445-4411-a2d0-c9e2618dd6bc","Type":"ContainerDied","Data":"dc3b37825d146cb808e7264a0b62e84dc6031e46bd732274f224c450da6148aa"} Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.293521 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6b7dd7f4b7-jwrwm"] Oct 13 13:00:10 crc kubenswrapper[4678]: E1013 13:00:10.293921 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" containerName="barbican-db-sync" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.293942 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" containerName="barbican-db-sync" Oct 13 13:00:10 crc kubenswrapper[4678]: E1013 13:00:10.293961 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="dnsmasq-dns" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.293970 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="dnsmasq-dns" Oct 13 13:00:10 crc kubenswrapper[4678]: E1013 13:00:10.293984 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="init" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.293993 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="init" Oct 13 13:00:10 crc kubenswrapper[4678]: E1013 13:00:10.294009 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0124058-8b2a-4a81-80ba-34fd9b93fd2b" containerName="collect-profiles" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.294016 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0124058-8b2a-4a81-80ba-34fd9b93fd2b" containerName="collect-profiles" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.294237 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4d3b9f-8933-4aa5-a660-90cbddd6fccb" containerName="dnsmasq-dns" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.294257 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0124058-8b2a-4a81-80ba-34fd9b93fd2b" containerName="collect-profiles" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.294280 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" containerName="barbican-db-sync" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.295388 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.305300 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dmgs8" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.305471 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.305597 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.326702 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-57bd8d65bb-fghbr"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.328175 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.331542 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342230 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342270 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-combined-ca-bundle\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342300 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42cvn\" (UniqueName: \"kubernetes.io/projected/b1fd2f17-a805-4985-afba-033768f4d77f-kube-api-access-42cvn\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342332 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1fd2f17-a805-4985-afba-033768f4d77f-logs\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342353 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data-custom\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.342481 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6b7dd7f4b7-jwrwm"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.389968 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57bd8d65bb-fghbr"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.456439 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467559 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data-custom\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467619 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-combined-ca-bundle\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467691 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42cvn\" (UniqueName: \"kubernetes.io/projected/b1fd2f17-a805-4985-afba-033768f4d77f-kube-api-access-42cvn\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467785 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1fd2f17-a805-4985-afba-033768f4d77f-logs\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467819 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data-custom\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467885 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-combined-ca-bundle\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.467926 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a9e7c5d-3070-4d64-b750-525ec923fc9c-logs\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.468208 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djnnn\" (UniqueName: \"kubernetes.io/projected/8a9e7c5d-3070-4d64-b750-525ec923fc9c-kube-api-access-djnnn\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.468252 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.471149 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1fd2f17-a805-4985-afba-033768f4d77f-logs\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.474903 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.480478 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-config-data-custom\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.480770 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1fd2f17-a805-4985-afba-033768f4d77f-combined-ca-bundle\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.498550 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.500536 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.508664 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42cvn\" (UniqueName: \"kubernetes.io/projected/b1fd2f17-a805-4985-afba-033768f4d77f-kube-api-access-42cvn\") pod \"barbican-worker-6b7dd7f4b7-jwrwm\" (UID: \"b1fd2f17-a805-4985-afba-033768f4d77f\") " pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.532870 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.571797 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.571933 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djnnn\" (UniqueName: \"kubernetes.io/projected/8a9e7c5d-3070-4d64-b750-525ec923fc9c-kube-api-access-djnnn\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.571963 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.578797 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.571990 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data-custom\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599515 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599546 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-combined-ca-bundle\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599578 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a9e7c5d-3070-4d64-b750-525ec923fc9c-logs\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599594 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599642 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwqhc\" (UniqueName: \"kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.599748 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.600698 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djnnn\" (UniqueName: \"kubernetes.io/projected/8a9e7c5d-3070-4d64-b750-525ec923fc9c-kube-api-access-djnnn\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.601433 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a9e7c5d-3070-4d64-b750-525ec923fc9c-logs\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.604357 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-combined-ca-bundle\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.604587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a9e7c5d-3070-4d64-b750-525ec923fc9c-config-data-custom\") pod \"barbican-keystone-listener-57bd8d65bb-fghbr\" (UID: \"8a9e7c5d-3070-4d64-b750-525ec923fc9c\") " pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.618107 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.620709 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.633075 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.654608 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.700160 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701073 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701659 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701740 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz55x\" (UniqueName: \"kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701788 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701823 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwqhc\" (UniqueName: \"kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701870 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701893 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701934 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.701975 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.702005 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.702032 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.702944 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.703577 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.706118 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.706570 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.708351 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.720929 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwqhc\" (UniqueName: \"kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc\") pod \"dnsmasq-dns-75c8ddd69c-zjnwp\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.803311 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.803380 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.803525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.803592 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.803634 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz55x\" (UniqueName: \"kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.808374 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.808854 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.813105 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.818444 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.823302 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz55x\" (UniqueName: \"kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x\") pod \"barbican-api-6d7ff444f6-wd4jp\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.949289 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:10 crc kubenswrapper[4678]: I1013 13:00:10.977013 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:12 crc kubenswrapper[4678]: I1013 13:00:12.905463 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6fd59c4d76-26t4f"] Oct 13 13:00:12 crc kubenswrapper[4678]: I1013 13:00:12.910135 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:12 crc kubenswrapper[4678]: I1013 13:00:12.912749 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 13 13:00:12 crc kubenswrapper[4678]: I1013 13:00:12.912925 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 13 13:00:12 crc kubenswrapper[4678]: I1013 13:00:12.984850 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fd59c4d76-26t4f"] Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058118 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-logs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzjpv\" (UniqueName: \"kubernetes.io/projected/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-kube-api-access-xzjpv\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058212 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-combined-ca-bundle\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data-custom\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058260 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-internal-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058275 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.058310 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-public-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.116617 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-logs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159509 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzjpv\" (UniqueName: \"kubernetes.io/projected/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-kube-api-access-xzjpv\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159559 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-combined-ca-bundle\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159592 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data-custom\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159612 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-internal-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159627 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.159662 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-public-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.160778 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-logs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.162884 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-94756fb8d-9h8cs" podUID="5bc24a68-de18-47d6-9573-67e978a05b05" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.150:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.150:8443: connect: connection refused" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.166110 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.166835 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-config-data-custom\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.167605 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-internal-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.168718 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-public-tls-certs\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.172730 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-combined-ca-bundle\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.199785 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzjpv\" (UniqueName: \"kubernetes.io/projected/aef9981e-9e19-4a3a-9f29-a82f4dfdc610-kube-api-access-xzjpv\") pod \"barbican-api-6fd59c4d76-26t4f\" (UID: \"aef9981e-9e19-4a3a-9f29-a82f4dfdc610\") " pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:13 crc kubenswrapper[4678]: I1013 13:00:13.226524 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.527166 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6mcn" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.617876 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.617931 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618035 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618089 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618205 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zhsn\" (UniqueName: \"kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618329 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data\") pod \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\" (UID: \"909be1eb-5445-4411-a2d0-c9e2618dd6bc\") " Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.618752 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/909be1eb-5445-4411-a2d0-c9e2618dd6bc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.625643 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts" (OuterVolumeSpecName: "scripts") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.640188 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn" (OuterVolumeSpecName: "kube-api-access-5zhsn") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "kube-api-access-5zhsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.642343 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.650500 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.686914 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data" (OuterVolumeSpecName: "config-data") pod "909be1eb-5445-4411-a2d0-c9e2618dd6bc" (UID: "909be1eb-5445-4411-a2d0-c9e2618dd6bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.720810 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.720841 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.720851 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.720859 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zhsn\" (UniqueName: \"kubernetes.io/projected/909be1eb-5445-4411-a2d0-c9e2618dd6bc-kube-api-access-5zhsn\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:16 crc kubenswrapper[4678]: I1013 13:00:16.720870 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/909be1eb-5445-4411-a2d0-c9e2618dd6bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.192936 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l6mcn" event={"ID":"909be1eb-5445-4411-a2d0-c9e2618dd6bc","Type":"ContainerDied","Data":"d074a156093305796441423107d27041fb59f228b237253c80641eba181ea36d"} Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.193569 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d074a156093305796441423107d27041fb59f228b237253c80641eba181ea36d" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.193186 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l6mcn" Oct 13 13:00:17 crc kubenswrapper[4678]: E1013 13:00:17.336478 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.586881 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fd59c4d76-26t4f"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.660265 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.708848 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57bd8d65bb-fghbr"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.747570 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.774084 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6b7dd7f4b7-jwrwm"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.849315 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:17 crc kubenswrapper[4678]: E1013 13:00:17.849711 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" containerName="cinder-db-sync" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.849729 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" containerName="cinder-db-sync" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.850565 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" containerName="cinder-db-sync" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.851665 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.854666 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6l4dl" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.854844 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.854948 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.855164 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.873151 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.884698 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.906226 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.909270 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.947645 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.948857 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.949063 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.949197 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7fbv\" (UniqueName: \"kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.949298 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.949410 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:17 crc kubenswrapper[4678]: I1013 13:00:17.949493 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.050620 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7fbv\" (UniqueName: \"kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.050954 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051003 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051085 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051114 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xw26\" (UniqueName: \"kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051187 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051214 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051254 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051274 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.051373 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.055517 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.057655 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.058342 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.060387 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.065624 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7fbv\" (UniqueName: \"kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv\") pod \"cinder-scheduler-0\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.093586 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.094998 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.099220 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.101892 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152291 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgjvh\" (UniqueName: \"kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152348 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152378 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152405 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152520 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xw26\" (UniqueName: \"kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.152573 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153285 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153483 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153556 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153604 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153690 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153784 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.153816 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.154008 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.154029 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.154513 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.173891 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xw26\" (UniqueName: \"kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26\") pod \"dnsmasq-dns-5784cf869f-chq27\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.204340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" event={"ID":"8a9e7c5d-3070-4d64-b750-525ec923fc9c","Type":"ContainerStarted","Data":"9d3852b70d68fb98bddfee8612f5bb1832bbe844b18db0e3ca41134212c3d82b"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.206194 4678 generic.go:334] "Generic (PLEG): container finished" podID="955aee00-7c8b-451f-ab36-3757d28a98e8" containerID="793b4f83bae69ce40912445dcb34fc6bc9e3b20a62780c2e48c460b86c857234" exitCode=0 Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.206343 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" event={"ID":"955aee00-7c8b-451f-ab36-3757d28a98e8","Type":"ContainerDied","Data":"793b4f83bae69ce40912445dcb34fc6bc9e3b20a62780c2e48c460b86c857234"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.206368 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" event={"ID":"955aee00-7c8b-451f-ab36-3757d28a98e8","Type":"ContainerStarted","Data":"7ea933c49ce7efd503704457cd90e6eb9de8f087692c41552521c09b92ad1ada"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.208512 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" event={"ID":"b1fd2f17-a805-4985-afba-033768f4d77f","Type":"ContainerStarted","Data":"ef9140570383fa5cbd30b1abd12ad3e08d84f59cb773ddb0fa04ca5cee022de1"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.210134 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.211265 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd59c4d76-26t4f" event={"ID":"aef9981e-9e19-4a3a-9f29-a82f4dfdc610","Type":"ContainerStarted","Data":"02aeac0c4c52a55179e53cdcd34a748cc4c4043f71bee3fdf17060fe21698976"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.211294 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd59c4d76-26t4f" event={"ID":"aef9981e-9e19-4a3a-9f29-a82f4dfdc610","Type":"ContainerStarted","Data":"666e0b35557511c3035914167f1ab73527ed17105214417b5dc565cd78907361"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.214481 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerStarted","Data":"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.214697 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="ceilometer-notification-agent" containerID="cri-o://3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841" gracePeriod=30 Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.214894 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.215001 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="proxy-httpd" containerID="cri-o://db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635" gracePeriod=30 Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.215152 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="sg-core" containerID="cri-o://96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed" gracePeriod=30 Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.241205 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerStarted","Data":"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.241245 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerStarted","Data":"0f60ecd0ec4db298828c330baaf86d5c9af7b6bcbbf5f30a39b109a71bbd4864"} Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.241996 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.242021 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.256100 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.256653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.256748 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.256841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.256969 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgjvh\" (UniqueName: \"kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.257073 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.257156 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.257536 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.257798 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.260273 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.260286 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.261256 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.271678 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.273824 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podStartSLOduration=8.27381001 podStartE2EDuration="8.27381001s" podCreationTimestamp="2025-10-13 13:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:18.267449553 +0000 UTC m=+1006.351987447" watchObservedRunningTime="2025-10-13 13:00:18.27381001 +0000 UTC m=+1006.358347894" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.276576 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgjvh\" (UniqueName: \"kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh\") pod \"cinder-api-0\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.280166 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.452424 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.635417 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.764003 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787466 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwqhc\" (UniqueName: \"kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787583 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787608 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787628 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787648 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.787724 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb\") pod \"955aee00-7c8b-451f-ab36-3757d28a98e8\" (UID: \"955aee00-7c8b-451f-ab36-3757d28a98e8\") " Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.806862 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc" (OuterVolumeSpecName: "kube-api-access-kwqhc") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "kube-api-access-kwqhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.825583 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.833542 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.842520 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config" (OuterVolumeSpecName: "config") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.853278 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.859299 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "955aee00-7c8b-451f-ab36-3757d28a98e8" (UID: "955aee00-7c8b-451f-ab36-3757d28a98e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890511 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwqhc\" (UniqueName: \"kubernetes.io/projected/955aee00-7c8b-451f-ab36-3757d28a98e8-kube-api-access-kwqhc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890552 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890568 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890581 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890594 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:18 crc kubenswrapper[4678]: I1013 13:00:18.890606 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/955aee00-7c8b-451f-ab36-3757d28a98e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.047098 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.070110 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:00:19 crc kubenswrapper[4678]: W1013 13:00:19.214758 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf8ebd25_714b_4864_ba75_551c20bf5390.slice/crio-1fdb1fa1fc721c08a9ede97fa5c447dda82697187e7e9f68b2a27c665318bfc9 WatchSource:0}: Error finding container 1fdb1fa1fc721c08a9ede97fa5c447dda82697187e7e9f68b2a27c665318bfc9: Status 404 returned error can't find the container with id 1fdb1fa1fc721c08a9ede97fa5c447dda82697187e7e9f68b2a27c665318bfc9 Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.250234 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerStarted","Data":"31b68876ea4c8219ea860420ac814479f27637b0dc4e07089c585dfb21217790"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.251785 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" event={"ID":"955aee00-7c8b-451f-ab36-3757d28a98e8","Type":"ContainerDied","Data":"7ea933c49ce7efd503704457cd90e6eb9de8f087692c41552521c09b92ad1ada"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.251817 4678 scope.go:117] "RemoveContainer" containerID="793b4f83bae69ce40912445dcb34fc6bc9e3b20a62780c2e48c460b86c857234" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.251917 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-zjnwp" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.255967 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd59c4d76-26t4f" event={"ID":"aef9981e-9e19-4a3a-9f29-a82f4dfdc610","Type":"ContainerStarted","Data":"be202f483a49a33675e5cf6184bbabb5ece620db03834e3b66b3202568983a2c"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.256406 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.256426 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.257726 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-chq27" event={"ID":"df8ebd25-714b-4864-ba75-551c20bf5390","Type":"ContainerStarted","Data":"1fdb1fa1fc721c08a9ede97fa5c447dda82697187e7e9f68b2a27c665318bfc9"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.261026 4678 generic.go:334] "Generic (PLEG): container finished" podID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerID="db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635" exitCode=0 Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.261060 4678 generic.go:334] "Generic (PLEG): container finished" podID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerID="96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed" exitCode=2 Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.261085 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerDied","Data":"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.261100 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerDied","Data":"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.262511 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerStarted","Data":"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1"} Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.283636 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6fd59c4d76-26t4f" podStartSLOduration=7.283618086 podStartE2EDuration="7.283618086s" podCreationTimestamp="2025-10-13 13:00:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:19.28224148 +0000 UTC m=+1007.366779364" watchObservedRunningTime="2025-10-13 13:00:19.283618086 +0000 UTC m=+1007.368155970" Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.327153 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:19 crc kubenswrapper[4678]: I1013 13:00:19.332573 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-zjnwp"] Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.047164 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.278703 4678 generic.go:334] "Generic (PLEG): container finished" podID="df8ebd25-714b-4864-ba75-551c20bf5390" containerID="0281e5471c7e348c85d88c0c81e48eff7380f30838df2998fe72eb41cc94a0f9" exitCode=0 Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.278759 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-chq27" event={"ID":"df8ebd25-714b-4864-ba75-551c20bf5390","Type":"ContainerDied","Data":"0281e5471c7e348c85d88c0c81e48eff7380f30838df2998fe72eb41cc94a0f9"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.282702 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerStarted","Data":"334d1eb76b660c6db57a3da342c7cf0f9b99af35cea0b751fe654a9e446ea235"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.286254 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" event={"ID":"8a9e7c5d-3070-4d64-b750-525ec923fc9c","Type":"ContainerStarted","Data":"c741c11ebe47f429686a66ab679bc41bb1486ee7a47c317c5b4c24d5698e8910"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.286307 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" event={"ID":"8a9e7c5d-3070-4d64-b750-525ec923fc9c","Type":"ContainerStarted","Data":"ca2418ad7cbd6b33003d58117574cd7d9771a426536e78ba1ef5bd8e52b3506d"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.293781 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" event={"ID":"b1fd2f17-a805-4985-afba-033768f4d77f","Type":"ContainerStarted","Data":"fccbb910e86bf20f2347b31400a0c7a9ae25670826faa5620c1f023b30dcdf8c"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.293828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" event={"ID":"b1fd2f17-a805-4985-afba-033768f4d77f","Type":"ContainerStarted","Data":"34cab857fa3c3de32e2379300bb305b28017113a8b2606440a177be693a60a8e"} Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.315855 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6b7dd7f4b7-jwrwm" podStartSLOduration=8.423344086 podStartE2EDuration="10.315810317s" podCreationTimestamp="2025-10-13 13:00:10 +0000 UTC" firstStartedPulling="2025-10-13 13:00:17.792627098 +0000 UTC m=+1005.877164982" lastFinishedPulling="2025-10-13 13:00:19.685093329 +0000 UTC m=+1007.769631213" observedRunningTime="2025-10-13 13:00:20.310257742 +0000 UTC m=+1008.394795646" watchObservedRunningTime="2025-10-13 13:00:20.315810317 +0000 UTC m=+1008.400348201" Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.325619 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-57bd8d65bb-fghbr" podStartSLOduration=8.306879056 podStartE2EDuration="10.325600254s" podCreationTimestamp="2025-10-13 13:00:10 +0000 UTC" firstStartedPulling="2025-10-13 13:00:17.663994949 +0000 UTC m=+1005.748532833" lastFinishedPulling="2025-10-13 13:00:19.682716147 +0000 UTC m=+1007.767254031" observedRunningTime="2025-10-13 13:00:20.325296286 +0000 UTC m=+1008.409834180" watchObservedRunningTime="2025-10-13 13:00:20.325600254 +0000 UTC m=+1008.410138138" Oct 13 13:00:20 crc kubenswrapper[4678]: I1013 13:00:20.613259 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="955aee00-7c8b-451f-ab36-3757d28a98e8" path="/var/lib/kubelet/pods/955aee00-7c8b-451f-ab36-3757d28a98e8/volumes" Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.306969 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-chq27" event={"ID":"df8ebd25-714b-4864-ba75-551c20bf5390","Type":"ContainerStarted","Data":"d4a1fc9b80b90962e4a82284d80916ad5913d1d1797e076e6b3d2813002e26f3"} Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.307449 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.311065 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerStarted","Data":"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a"} Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.311120 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerStarted","Data":"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee"} Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.311142 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api-log" containerID="cri-o://da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee" gracePeriod=30 Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.311166 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.311200 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api" containerID="cri-o://d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a" gracePeriod=30 Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.313482 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerStarted","Data":"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f"} Oct 13 13:00:21 crc kubenswrapper[4678]: I1013 13:00:21.350117 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-chq27" podStartSLOduration=4.350096365 podStartE2EDuration="4.350096365s" podCreationTimestamp="2025-10-13 13:00:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:21.339529298 +0000 UTC m=+1009.424067192" watchObservedRunningTime="2025-10-13 13:00:21.350096365 +0000 UTC m=+1009.434634249" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.035174 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.054696 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.054678307 podStartE2EDuration="4.054678307s" podCreationTimestamp="2025-10-13 13:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:21.369637176 +0000 UTC m=+1009.454175070" watchObservedRunningTime="2025-10-13 13:00:22.054678307 +0000 UTC m=+1010.139216191" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.145299 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259073 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259174 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259191 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259214 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259290 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259363 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfvm5\" (UniqueName: \"kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.259410 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd\") pod \"59f4d1c3-4733-46a9-8502-d03770bc285d\" (UID: \"59f4d1c3-4733-46a9-8502-d03770bc285d\") " Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.260017 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.260182 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.264994 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5" (OuterVolumeSpecName: "kube-api-access-xfvm5") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "kube-api-access-xfvm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.265563 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts" (OuterVolumeSpecName: "scripts") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.303409 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.333148 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.339511 4678 generic.go:334] "Generic (PLEG): container finished" podID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerID="3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841" exitCode=0 Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.339581 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerDied","Data":"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841"} Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.339587 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.339610 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59f4d1c3-4733-46a9-8502-d03770bc285d","Type":"ContainerDied","Data":"b5373016d9bd2599884ebd9ad166edc3491a32a475f99e283960b1f0669caac2"} Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.339627 4678 scope.go:117] "RemoveContainer" containerID="db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.349493 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerDied","Data":"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee"} Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.349538 4678 generic.go:334] "Generic (PLEG): container finished" podID="51a84090-775e-432c-91e0-f142e46147da" containerID="da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee" exitCode=143 Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.352416 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data" (OuterVolumeSpecName: "config-data") pod "59f4d1c3-4733-46a9-8502-d03770bc285d" (UID: "59f4d1c3-4733-46a9-8502-d03770bc285d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.353710 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerStarted","Data":"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1"} Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369197 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369238 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369248 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369257 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369266 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59f4d1c3-4733-46a9-8502-d03770bc285d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369273 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59f4d1c3-4733-46a9-8502-d03770bc285d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369281 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfvm5\" (UniqueName: \"kubernetes.io/projected/59f4d1c3-4733-46a9-8502-d03770bc285d-kube-api-access-xfvm5\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.369449 4678 scope.go:117] "RemoveContainer" containerID="96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.376771 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.086726048 podStartE2EDuration="5.376754042s" podCreationTimestamp="2025-10-13 13:00:17 +0000 UTC" firstStartedPulling="2025-10-13 13:00:18.770288542 +0000 UTC m=+1006.854826426" lastFinishedPulling="2025-10-13 13:00:20.060316536 +0000 UTC m=+1008.144854420" observedRunningTime="2025-10-13 13:00:22.374601686 +0000 UTC m=+1010.459139590" watchObservedRunningTime="2025-10-13 13:00:22.376754042 +0000 UTC m=+1010.461291926" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.406321 4678 scope.go:117] "RemoveContainer" containerID="3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.425238 4678 scope.go:117] "RemoveContainer" containerID="db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635" Oct 13 13:00:22 crc kubenswrapper[4678]: E1013 13:00:22.425744 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635\": container with ID starting with db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635 not found: ID does not exist" containerID="db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.425782 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635"} err="failed to get container status \"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635\": rpc error: code = NotFound desc = could not find container \"db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635\": container with ID starting with db9b88bce56f42a30d9b614a24f1c440cb5c28cb0c0ea70f353cacea79cfe635 not found: ID does not exist" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.425805 4678 scope.go:117] "RemoveContainer" containerID="96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed" Oct 13 13:00:22 crc kubenswrapper[4678]: E1013 13:00:22.427320 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed\": container with ID starting with 96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed not found: ID does not exist" containerID="96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.427391 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed"} err="failed to get container status \"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed\": rpc error: code = NotFound desc = could not find container \"96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed\": container with ID starting with 96f34ab500d81f2d7d3e4ace26b7ad2518dcadad0ed683f2830be844d29092ed not found: ID does not exist" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.427424 4678 scope.go:117] "RemoveContainer" containerID="3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841" Oct 13 13:00:22 crc kubenswrapper[4678]: E1013 13:00:22.428231 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841\": container with ID starting with 3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841 not found: ID does not exist" containerID="3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.428257 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841"} err="failed to get container status \"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841\": rpc error: code = NotFound desc = could not find container \"3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841\": container with ID starting with 3f4300f75c61608263ec0fa7c474f517c6f7c8517bae527bf619ea912ad22841 not found: ID does not exist" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.976848 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.981827 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f64664c65-vk8pt" Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.985864 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:22 crc kubenswrapper[4678]: I1013 13:00:22.998550 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999276 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="955aee00-7c8b-451f-ab36-3757d28a98e8" containerName="init" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999296 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="955aee00-7c8b-451f-ab36-3757d28a98e8" containerName="init" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999307 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999313 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999324 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="ceilometer-notification-agent" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999330 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="ceilometer-notification-agent" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999346 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="sg-core" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999353 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="sg-core" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999361 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="proxy-httpd" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999366 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="proxy-httpd" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:22.999375 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon-log" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999381 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon-log" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999538 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="ceilometer-notification-agent" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999551 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon-log" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999562 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerName="horizon" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999571 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="proxy-httpd" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999579 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" containerName="sg-core" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:22.999592 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="955aee00-7c8b-451f-ab36-3757d28a98e8" containerName="init" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.007844 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.010710 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.011922 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.022625 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.024164 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090573 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key\") pod \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090706 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key\") pod \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090728 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs\") pod \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090747 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts\") pod \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090814 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data\") pod \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090841 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts\") pod \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090879 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data\") pod \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090906 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb2nj\" (UniqueName: \"kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj\") pod \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\" (UID: \"9b57c23d-0e94-437e-969a-e9c9f2e56d3d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.090988 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pztjx\" (UniqueName: \"kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx\") pod \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091008 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs\") pod \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\" (UID: \"8d9788bd-6bed-4e7d-89c3-950a068a2c6d\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091251 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091312 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091333 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091368 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091394 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk4x6\" (UniqueName: \"kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091443 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.091456 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.096103 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs" (OuterVolumeSpecName: "logs") pod "8d9788bd-6bed-4e7d-89c3-950a068a2c6d" (UID: "8d9788bd-6bed-4e7d-89c3-950a068a2c6d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.096137 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs" (OuterVolumeSpecName: "logs") pod "9b57c23d-0e94-437e-969a-e9c9f2e56d3d" (UID: "9b57c23d-0e94-437e-969a-e9c9f2e56d3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.096257 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9b57c23d-0e94-437e-969a-e9c9f2e56d3d" (UID: "9b57c23d-0e94-437e-969a-e9c9f2e56d3d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.097472 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx" (OuterVolumeSpecName: "kube-api-access-pztjx") pod "8d9788bd-6bed-4e7d-89c3-950a068a2c6d" (UID: "8d9788bd-6bed-4e7d-89c3-950a068a2c6d"). InnerVolumeSpecName "kube-api-access-pztjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.100152 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8d9788bd-6bed-4e7d-89c3-950a068a2c6d" (UID: "8d9788bd-6bed-4e7d-89c3-950a068a2c6d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.101652 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj" (OuterVolumeSpecName: "kube-api-access-xb2nj") pod "9b57c23d-0e94-437e-969a-e9c9f2e56d3d" (UID: "9b57c23d-0e94-437e-969a-e9c9f2e56d3d"). InnerVolumeSpecName "kube-api-access-xb2nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.132561 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts" (OuterVolumeSpecName: "scripts") pod "8d9788bd-6bed-4e7d-89c3-950a068a2c6d" (UID: "8d9788bd-6bed-4e7d-89c3-950a068a2c6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.135695 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.139007 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts" (OuterVolumeSpecName: "scripts") pod "9b57c23d-0e94-437e-969a-e9c9f2e56d3d" (UID: "9b57c23d-0e94-437e-969a-e9c9f2e56d3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.140191 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data" (OuterVolumeSpecName: "config-data") pod "8d9788bd-6bed-4e7d-89c3-950a068a2c6d" (UID: "8d9788bd-6bed-4e7d-89c3-950a068a2c6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.145412 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data" (OuterVolumeSpecName: "config-data") pod "9b57c23d-0e94-437e-969a-e9c9f2e56d3d" (UID: "9b57c23d-0e94-437e-969a-e9c9f2e56d3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192474 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk4x6\" (UniqueName: \"kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192550 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192567 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192606 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192654 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192675 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192711 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192759 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192771 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192779 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192789 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192797 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192805 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192814 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb2nj\" (UniqueName: \"kubernetes.io/projected/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-kube-api-access-xb2nj\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192824 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pztjx\" (UniqueName: \"kubernetes.io/projected/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-kube-api-access-pztjx\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192832 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d9788bd-6bed-4e7d-89c3-950a068a2c6d-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.192841 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9b57c23d-0e94-437e-969a-e9c9f2e56d3d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.193971 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.194207 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.197004 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.197358 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.198540 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.198986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.209880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk4x6\" (UniqueName: \"kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6\") pod \"ceilometer-0\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.211652 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.293732 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs\") pod \"9dd963a8-d4b0-4a3b-b134-663baf672b41\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.293836 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data\") pod \"9dd963a8-d4b0-4a3b-b134-663baf672b41\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.293899 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f24tx\" (UniqueName: \"kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx\") pod \"9dd963a8-d4b0-4a3b-b134-663baf672b41\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.293970 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key\") pod \"9dd963a8-d4b0-4a3b-b134-663baf672b41\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.294118 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts\") pod \"9dd963a8-d4b0-4a3b-b134-663baf672b41\" (UID: \"9dd963a8-d4b0-4a3b-b134-663baf672b41\") " Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.295375 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs" (OuterVolumeSpecName: "logs") pod "9dd963a8-d4b0-4a3b-b134-663baf672b41" (UID: "9dd963a8-d4b0-4a3b-b134-663baf672b41"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.299326 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9dd963a8-d4b0-4a3b-b134-663baf672b41" (UID: "9dd963a8-d4b0-4a3b-b134-663baf672b41"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.307011 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx" (OuterVolumeSpecName: "kube-api-access-f24tx") pod "9dd963a8-d4b0-4a3b-b134-663baf672b41" (UID: "9dd963a8-d4b0-4a3b-b134-663baf672b41"). InnerVolumeSpecName "kube-api-access-f24tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.315030 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data" (OuterVolumeSpecName: "config-data") pod "9dd963a8-d4b0-4a3b-b134-663baf672b41" (UID: "9dd963a8-d4b0-4a3b-b134-663baf672b41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.317334 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts" (OuterVolumeSpecName: "scripts") pod "9dd963a8-d4b0-4a3b-b134-663baf672b41" (UID: "9dd963a8-d4b0-4a3b-b134-663baf672b41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.341538 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384326 4678 generic.go:334] "Generic (PLEG): container finished" podID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerID="236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384678 4678 generic.go:334] "Generic (PLEG): container finished" podID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" containerID="5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384509 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f64664c65-vk8pt" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384515 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerDied","Data":"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384962 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerDied","Data":"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.384995 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f64664c65-vk8pt" event={"ID":"8d9788bd-6bed-4e7d-89c3-950a068a2c6d","Type":"ContainerDied","Data":"f36212f0bb412201b6722469b788aa6e63f07f8c301dd662398246671ecd1491"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.385021 4678 scope.go:117] "RemoveContainer" containerID="236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.396325 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.397033 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd963a8-d4b0-4a3b-b134-663baf672b41-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.397250 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9dd963a8-d4b0-4a3b-b134-663baf672b41-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.397469 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f24tx\" (UniqueName: \"kubernetes.io/projected/9dd963a8-d4b0-4a3b-b134-663baf672b41-kube-api-access-f24tx\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.397647 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9dd963a8-d4b0-4a3b-b134-663baf672b41-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.400674 4678 generic.go:334] "Generic (PLEG): container finished" podID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerID="4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.400834 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6bb9d9cd79-jbbgw" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.401043 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerDied","Data":"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.401312 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerDied","Data":"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.400837 4678 generic.go:334] "Generic (PLEG): container finished" podID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerID="dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.401673 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6bb9d9cd79-jbbgw" event={"ID":"9dd963a8-d4b0-4a3b-b134-663baf672b41","Type":"ContainerDied","Data":"9e4775e82f993f29874255d7458f16590fa0324fd7e6675522ceb7c31b541a74"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.410475 4678 generic.go:334] "Generic (PLEG): container finished" podID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerID="a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.410520 4678 generic.go:334] "Generic (PLEG): container finished" podID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerID="a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" exitCode=137 Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.411534 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bfb45cccf-j6vzn" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.411864 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerDied","Data":"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.411948 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerDied","Data":"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.411972 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bfb45cccf-j6vzn" event={"ID":"9b57c23d-0e94-437e-969a-e9c9f2e56d3d","Type":"ContainerDied","Data":"2518d676b50fc1c2732a42c26d1b9df5b4601ea707d7e1c2eb06a18ee4650ebb"} Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.465284 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.506307 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f64664c65-vk8pt"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.540395 4678 scope.go:117] "RemoveContainer" containerID="5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.545120 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.574426 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6bb9d9cd79-jbbgw"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.584098 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.590492 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5bfb45cccf-j6vzn"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.597589 4678 scope.go:117] "RemoveContainer" containerID="236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:23.598073 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a\": container with ID starting with 236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a not found: ID does not exist" containerID="236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598100 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a"} err="failed to get container status \"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a\": rpc error: code = NotFound desc = could not find container \"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a\": container with ID starting with 236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598122 4678 scope.go:117] "RemoveContainer" containerID="5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:23.598336 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba\": container with ID starting with 5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba not found: ID does not exist" containerID="5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598352 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba"} err="failed to get container status \"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba\": rpc error: code = NotFound desc = could not find container \"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba\": container with ID starting with 5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598364 4678 scope.go:117] "RemoveContainer" containerID="236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598511 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a"} err="failed to get container status \"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a\": rpc error: code = NotFound desc = could not find container \"236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a\": container with ID starting with 236e88d6d2cee7caf321a92b50f227250b3b6f8128337b218a709d8e4537c25a not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598525 4678 scope.go:117] "RemoveContainer" containerID="5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598666 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba"} err="failed to get container status \"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba\": rpc error: code = NotFound desc = could not find container \"5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba\": container with ID starting with 5a94728bda23079b2a0b1ce40c83fe2707505e7c9997fa495c8f5103bc3155ba not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.598679 4678 scope.go:117] "RemoveContainer" containerID="4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.797391 4678 scope.go:117] "RemoveContainer" containerID="dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.813282 4678 scope.go:117] "RemoveContainer" containerID="4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:23.813829 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52\": container with ID starting with 4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52 not found: ID does not exist" containerID="4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.813948 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52"} err="failed to get container status \"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52\": rpc error: code = NotFound desc = could not find container \"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52\": container with ID starting with 4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52 not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.813990 4678 scope.go:117] "RemoveContainer" containerID="dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" Oct 13 13:00:23 crc kubenswrapper[4678]: E1013 13:00:23.814429 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7\": container with ID starting with dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7 not found: ID does not exist" containerID="dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.814458 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7"} err="failed to get container status \"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7\": rpc error: code = NotFound desc = could not find container \"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7\": container with ID starting with dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7 not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.814501 4678 scope.go:117] "RemoveContainer" containerID="4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.814835 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52"} err="failed to get container status \"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52\": rpc error: code = NotFound desc = could not find container \"4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52\": container with ID starting with 4409ec4117b58463fb4a938b02708e158629bf33163820634df0a5e511d4ce52 not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.814874 4678 scope.go:117] "RemoveContainer" containerID="dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.815211 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7"} err="failed to get container status \"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7\": rpc error: code = NotFound desc = could not find container \"dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7\": container with ID starting with dc7560b3718efc8d7130f707157d5508469dc8f4b74838788fef5633f14148a7 not found: ID does not exist" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.815231 4678 scope.go:117] "RemoveContainer" containerID="a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.952389 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:23 crc kubenswrapper[4678]: I1013 13:00:23.984486 4678 scope.go:117] "RemoveContainer" containerID="a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.070846 4678 scope.go:117] "RemoveContainer" containerID="a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" Oct 13 13:00:24 crc kubenswrapper[4678]: E1013 13:00:24.071544 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139\": container with ID starting with a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139 not found: ID does not exist" containerID="a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.071594 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139"} err="failed to get container status \"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139\": rpc error: code = NotFound desc = could not find container \"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139\": container with ID starting with a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139 not found: ID does not exist" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.071627 4678 scope.go:117] "RemoveContainer" containerID="a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" Oct 13 13:00:24 crc kubenswrapper[4678]: E1013 13:00:24.072019 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a\": container with ID starting with a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a not found: ID does not exist" containerID="a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.072046 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a"} err="failed to get container status \"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a\": rpc error: code = NotFound desc = could not find container \"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a\": container with ID starting with a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a not found: ID does not exist" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.072099 4678 scope.go:117] "RemoveContainer" containerID="a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.072452 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139"} err="failed to get container status \"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139\": rpc error: code = NotFound desc = could not find container \"a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139\": container with ID starting with a4c09bbaa5d620ba13a44956958c7617e7480e08f95edfd73421d41a6fdfe139 not found: ID does not exist" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.072479 4678 scope.go:117] "RemoveContainer" containerID="a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.072831 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a"} err="failed to get container status \"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a\": rpc error: code = NotFound desc = could not find container \"a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a\": container with ID starting with a7b8cb46a870e3807b9d4f44203838bbbae4104105c74d749700fcc04fef819a not found: ID does not exist" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.421420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerStarted","Data":"09019e02c7f58eceef4f3a5d9529873c5ef3c8842eff06f2aec6ce7f967996c8"} Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.604673 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59f4d1c3-4733-46a9-8502-d03770bc285d" path="/var/lib/kubelet/pods/59f4d1c3-4733-46a9-8502-d03770bc285d/volumes" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.605659 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d9788bd-6bed-4e7d-89c3-950a068a2c6d" path="/var/lib/kubelet/pods/8d9788bd-6bed-4e7d-89c3-950a068a2c6d/volumes" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.607187 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" path="/var/lib/kubelet/pods/9b57c23d-0e94-437e-969a-e9c9f2e56d3d/volumes" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.608083 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" path="/var/lib/kubelet/pods/9dd963a8-d4b0-4a3b-b134-663baf672b41/volumes" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.647957 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.688445 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-dcbfc8897-x7vwx" Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.762960 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.763551 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c7b867ddd-k78g5" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-api" containerID="cri-o://c72e762a63fc27cfa9694add9b2859136f37527127fc566769862abbb1e9dd6b" gracePeriod=30 Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.763963 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c7b867ddd-k78g5" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-httpd" containerID="cri-o://981a5d7b748bc3ac9dd71d235f47c8d951188cc2a02481a5101f52897a4effc8" gracePeriod=30 Oct 13 13:00:24 crc kubenswrapper[4678]: I1013 13:00:24.943224 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.007339 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.164655 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fd59c4d76-26t4f" Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.233430 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.233688 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" containerID="cri-o://d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898" gracePeriod=30 Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.234238 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" containerID="cri-o://5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1" gracePeriod=30 Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.242247 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.242351 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.461893 4678 generic.go:334] "Generic (PLEG): container finished" podID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerID="981a5d7b748bc3ac9dd71d235f47c8d951188cc2a02481a5101f52897a4effc8" exitCode=0 Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.461950 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerDied","Data":"981a5d7b748bc3ac9dd71d235f47c8d951188cc2a02481a5101f52897a4effc8"} Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.465536 4678 generic.go:334] "Generic (PLEG): container finished" podID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerID="d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898" exitCode=143 Oct 13 13:00:25 crc kubenswrapper[4678]: I1013 13:00:25.466176 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerDied","Data":"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898"} Oct 13 13:00:26 crc kubenswrapper[4678]: I1013 13:00:26.474687 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerStarted","Data":"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68"} Oct 13 13:00:26 crc kubenswrapper[4678]: I1013 13:00:26.474726 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerStarted","Data":"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74"} Oct 13 13:00:26 crc kubenswrapper[4678]: I1013 13:00:26.915313 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:26 crc kubenswrapper[4678]: I1013 13:00:26.982337 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-94756fb8d-9h8cs" Oct 13 13:00:27 crc kubenswrapper[4678]: I1013 13:00:27.057541 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 13:00:27 crc kubenswrapper[4678]: I1013 13:00:27.484813 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon-log" containerID="cri-o://4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71" gracePeriod=30 Oct 13 13:00:27 crc kubenswrapper[4678]: I1013 13:00:27.486411 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerStarted","Data":"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3"} Oct 13 13:00:27 crc kubenswrapper[4678]: I1013 13:00:27.487078 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" containerID="cri-o://00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1" gracePeriod=30 Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.279190 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.333522 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.333746 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="dnsmasq-dns" containerID="cri-o://0ed8f5ca616bc1e655ccf3d1bc269a84f3ea4b5e468b876edb9134acb7cb6fea" gracePeriod=10 Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.444582 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.507171 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.520004 4678 generic.go:334] "Generic (PLEG): container finished" podID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerID="0ed8f5ca616bc1e655ccf3d1bc269a84f3ea4b5e468b876edb9134acb7cb6fea" exitCode=0 Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.520251 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="cinder-scheduler" containerID="cri-o://086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f" gracePeriod=30 Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.520534 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" event={"ID":"47e524d5-b1ac-4559-af3a-b56fe79105a3","Type":"ContainerDied","Data":"0ed8f5ca616bc1e655ccf3d1bc269a84f3ea4b5e468b876edb9134acb7cb6fea"} Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.520860 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="probe" containerID="cri-o://8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1" gracePeriod=30 Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.867986 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919419 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919484 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919527 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919558 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919578 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.919674 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmrzp\" (UniqueName: \"kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp\") pod \"47e524d5-b1ac-4559-af3a-b56fe79105a3\" (UID: \"47e524d5-b1ac-4559-af3a-b56fe79105a3\") " Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.934351 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp" (OuterVolumeSpecName: "kube-api-access-dmrzp") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "kube-api-access-dmrzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.973576 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.979712 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config" (OuterVolumeSpecName: "config") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.981702 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.982691 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:28 crc kubenswrapper[4678]: I1013 13:00:28.994732 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "47e524d5-b1ac-4559-af3a-b56fe79105a3" (UID: "47e524d5-b1ac-4559-af3a-b56fe79105a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021423 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021655 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021754 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021821 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021880 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47e524d5-b1ac-4559-af3a-b56fe79105a3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.021931 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmrzp\" (UniqueName: \"kubernetes.io/projected/47e524d5-b1ac-4559-af3a-b56fe79105a3-kube-api-access-dmrzp\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.537849 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" event={"ID":"47e524d5-b1ac-4559-af3a-b56fe79105a3","Type":"ContainerDied","Data":"decb5a13bd5b1005b9bcd4b46913609f990798b73577cae84278638fce2399b4"} Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.538199 4678 scope.go:117] "RemoveContainer" containerID="0ed8f5ca616bc1e655ccf3d1bc269a84f3ea4b5e468b876edb9134acb7cb6fea" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.538367 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-k9xl9" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.545173 4678 generic.go:334] "Generic (PLEG): container finished" podID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerID="8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1" exitCode=0 Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.545241 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerDied","Data":"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1"} Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.561235 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerStarted","Data":"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0"} Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.561478 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.566591 4678 scope.go:117] "RemoveContainer" containerID="a96e400fd8cdafdb684c1a8687da680a24818d6b91fa0be598c91762dfcc4ae4" Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.580305 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.586588 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-k9xl9"] Oct 13 13:00:29 crc kubenswrapper[4678]: I1013 13:00:29.609533 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.016819973 podStartE2EDuration="7.609517941s" podCreationTimestamp="2025-10-13 13:00:22 +0000 UTC" firstStartedPulling="2025-10-13 13:00:23.995492175 +0000 UTC m=+1012.080030049" lastFinishedPulling="2025-10-13 13:00:28.588190143 +0000 UTC m=+1016.672728017" observedRunningTime="2025-10-13 13:00:29.604426727 +0000 UTC m=+1017.688964611" watchObservedRunningTime="2025-10-13 13:00:29.609517941 +0000 UTC m=+1017.694055825" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.221248 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.571552 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerDied","Data":"c72e762a63fc27cfa9694add9b2859136f37527127fc566769862abbb1e9dd6b"} Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.571571 4678 generic.go:334] "Generic (PLEG): container finished" podID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerID="c72e762a63fc27cfa9694add9b2859136f37527127fc566769862abbb1e9dd6b" exitCode=0 Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.604708 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" path="/var/lib/kubelet/pods/47e524d5-b1ac-4559-af3a-b56fe79105a3/volumes" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.639743 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-694bd84568-p2blx" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.653062 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-694bd84568-p2blx" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.679113 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:53746->10.217.0.163:9311: read: connection reset by peer" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.679189 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:53754->10.217.0.163:9311: read: connection reset by peer" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.868347 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.960068 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config\") pod \"79788fad-4cc8-442b-9ba6-698e1a05a472\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.960139 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs\") pod \"79788fad-4cc8-442b-9ba6-698e1a05a472\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.960230 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle\") pod \"79788fad-4cc8-442b-9ba6-698e1a05a472\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.960314 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25dwj\" (UniqueName: \"kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj\") pod \"79788fad-4cc8-442b-9ba6-698e1a05a472\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.960399 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config\") pod \"79788fad-4cc8-442b-9ba6-698e1a05a472\" (UID: \"79788fad-4cc8-442b-9ba6-698e1a05a472\") " Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.982444 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "79788fad-4cc8-442b-9ba6-698e1a05a472" (UID: "79788fad-4cc8-442b-9ba6-698e1a05a472"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:30 crc kubenswrapper[4678]: I1013 13:00:30.985318 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj" (OuterVolumeSpecName: "kube-api-access-25dwj") pod "79788fad-4cc8-442b-9ba6-698e1a05a472" (UID: "79788fad-4cc8-442b-9ba6-698e1a05a472"). InnerVolumeSpecName "kube-api-access-25dwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.012170 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79788fad-4cc8-442b-9ba6-698e1a05a472" (UID: "79788fad-4cc8-442b-9ba6-698e1a05a472"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.012626 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config" (OuterVolumeSpecName: "config") pod "79788fad-4cc8-442b-9ba6-698e1a05a472" (UID: "79788fad-4cc8-442b-9ba6-698e1a05a472"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.031912 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "79788fad-4cc8-442b-9ba6-698e1a05a472" (UID: "79788fad-4cc8-442b-9ba6-698e1a05a472"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.063603 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25dwj\" (UniqueName: \"kubernetes.io/projected/79788fad-4cc8-442b-9ba6-698e1a05a472-kube-api-access-25dwj\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.063635 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.063645 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.063654 4678 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.063664 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79788fad-4cc8-442b-9ba6-698e1a05a472-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.071799 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.169279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data\") pod \"d2954d85-0d57-4f49-91f6-971b0c9005b5\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.169377 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs\") pod \"d2954d85-0d57-4f49-91f6-971b0c9005b5\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.169529 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom\") pod \"d2954d85-0d57-4f49-91f6-971b0c9005b5\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.169580 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz55x\" (UniqueName: \"kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x\") pod \"d2954d85-0d57-4f49-91f6-971b0c9005b5\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.169612 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle\") pod \"d2954d85-0d57-4f49-91f6-971b0c9005b5\" (UID: \"d2954d85-0d57-4f49-91f6-971b0c9005b5\") " Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.170800 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs" (OuterVolumeSpecName: "logs") pod "d2954d85-0d57-4f49-91f6-971b0c9005b5" (UID: "d2954d85-0d57-4f49-91f6-971b0c9005b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.174618 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x" (OuterVolumeSpecName: "kube-api-access-tz55x") pod "d2954d85-0d57-4f49-91f6-971b0c9005b5" (UID: "d2954d85-0d57-4f49-91f6-971b0c9005b5"). InnerVolumeSpecName "kube-api-access-tz55x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.175178 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d2954d85-0d57-4f49-91f6-971b0c9005b5" (UID: "d2954d85-0d57-4f49-91f6-971b0c9005b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.201526 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2954d85-0d57-4f49-91f6-971b0c9005b5" (UID: "d2954d85-0d57-4f49-91f6-971b0c9005b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.260542 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data" (OuterVolumeSpecName: "config-data") pod "d2954d85-0d57-4f49-91f6-971b0c9005b5" (UID: "d2954d85-0d57-4f49-91f6-971b0c9005b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.271769 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.272117 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz55x\" (UniqueName: \"kubernetes.io/projected/d2954d85-0d57-4f49-91f6-971b0c9005b5-kube-api-access-tz55x\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.272221 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.272277 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2954d85-0d57-4f49-91f6-971b0c9005b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.272337 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2954d85-0d57-4f49-91f6-971b0c9005b5-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.588298 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7b867ddd-k78g5" event={"ID":"79788fad-4cc8-442b-9ba6-698e1a05a472","Type":"ContainerDied","Data":"81ef72ca8bdc0f7c4c8c36e1b85ba85c4d0ca47b4defac9c5f057baff4fabc7c"} Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.588349 4678 scope.go:117] "RemoveContainer" containerID="981a5d7b748bc3ac9dd71d235f47c8d951188cc2a02481a5101f52897a4effc8" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.589831 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7b867ddd-k78g5" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.591634 4678 generic.go:334] "Generic (PLEG): container finished" podID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerID="5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1" exitCode=0 Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.591709 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerDied","Data":"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1"} Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.591730 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d7ff444f6-wd4jp" event={"ID":"d2954d85-0d57-4f49-91f6-971b0c9005b5","Type":"ContainerDied","Data":"0f60ecd0ec4db298828c330baaf86d5c9af7b6bcbbf5f30a39b109a71bbd4864"} Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.591778 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d7ff444f6-wd4jp" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.595085 4678 generic.go:334] "Generic (PLEG): container finished" podID="a4901114-e36f-4b5b-8d84-d64f02997156" containerID="00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1" exitCode=0 Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.595555 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerDied","Data":"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1"} Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.640871 4678 scope.go:117] "RemoveContainer" containerID="c72e762a63fc27cfa9694add9b2859136f37527127fc566769862abbb1e9dd6b" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.659070 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.670337 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6d7ff444f6-wd4jp"] Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.677653 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.693261 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c7b867ddd-k78g5"] Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.715728 4678 scope.go:117] "RemoveContainer" containerID="5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.743450 4678 scope.go:117] "RemoveContainer" containerID="d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.796673 4678 scope.go:117] "RemoveContainer" containerID="5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1" Oct 13 13:00:31 crc kubenswrapper[4678]: E1013 13:00:31.797163 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1\": container with ID starting with 5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1 not found: ID does not exist" containerID="5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.797209 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1"} err="failed to get container status \"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1\": rpc error: code = NotFound desc = could not find container \"5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1\": container with ID starting with 5cc745a166b1e0cae10217839d0e627dc5b7bca348562c55b05c1788d04e76d1 not found: ID does not exist" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.797237 4678 scope.go:117] "RemoveContainer" containerID="d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898" Oct 13 13:00:31 crc kubenswrapper[4678]: E1013 13:00:31.797578 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898\": container with ID starting with d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898 not found: ID does not exist" containerID="d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898" Oct 13 13:00:31 crc kubenswrapper[4678]: I1013 13:00:31.797649 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898"} err="failed to get container status \"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898\": rpc error: code = NotFound desc = could not find container \"d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898\": container with ID starting with d59e1a5efb2c531f514981029ad2a392da318ca1c7d808dcf0b52c00e9ede898 not found: ID does not exist" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.132480 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190102 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190185 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7fbv\" (UniqueName: \"kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190235 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190453 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190529 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.190632 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle\") pod \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\" (UID: \"84ac69b4-d0c9-4972-90c4-8232dd9986b1\") " Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.192426 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.195357 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.198145 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts" (OuterVolumeSpecName: "scripts") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.199651 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv" (OuterVolumeSpecName: "kube-api-access-g7fbv") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "kube-api-access-g7fbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.248183 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.288866 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data" (OuterVolumeSpecName: "config-data") pod "84ac69b4-d0c9-4972-90c4-8232dd9986b1" (UID: "84ac69b4-d0c9-4972-90c4-8232dd9986b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293217 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/84ac69b4-d0c9-4972-90c4-8232dd9986b1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293261 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293276 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293287 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7fbv\" (UniqueName: \"kubernetes.io/projected/84ac69b4-d0c9-4972-90c4-8232dd9986b1-kube-api-access-g7fbv\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293301 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.293312 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/84ac69b4-d0c9-4972-90c4-8232dd9986b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.624921 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" path="/var/lib/kubelet/pods/79788fad-4cc8-442b-9ba6-698e1a05a472/volumes" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.626311 4678 generic.go:334] "Generic (PLEG): container finished" podID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerID="086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f" exitCode=0 Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.626529 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.637416 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" path="/var/lib/kubelet/pods/d2954d85-0d57-4f49-91f6-971b0c9005b5/volumes" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.638692 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerDied","Data":"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f"} Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.638724 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"84ac69b4-d0c9-4972-90c4-8232dd9986b1","Type":"ContainerDied","Data":"31b68876ea4c8219ea860420ac814479f27637b0dc4e07089c585dfb21217790"} Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.638745 4678 scope.go:117] "RemoveContainer" containerID="8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.661096 4678 scope.go:117] "RemoveContainer" containerID="086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.677203 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.685377 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700297 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700617 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="probe" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700629 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="probe" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700644 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700650 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700661 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-httpd" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700915 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-httpd" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700936 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700942 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700952 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700958 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700968 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="init" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700974 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="init" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700984 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.700989 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.700998 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="dnsmasq-dns" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701004 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="dnsmasq-dns" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.701026 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701032 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.701043 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-api" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701063 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-api" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.701079 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="cinder-scheduler" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701085 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="cinder-scheduler" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.701094 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701100 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701776 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701793 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701807 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701818 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="probe" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701827 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" containerName="cinder-scheduler" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701838 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e524d5-b1ac-4559-af3a-b56fe79105a3" containerName="dnsmasq-dns" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701848 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701856 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd963a8-d4b0-4a3b-b134-663baf672b41" containerName="horizon-log" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701865 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-api" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701871 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b57c23d-0e94-437e-969a-e9c9f2e56d3d" containerName="horizon" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.701882 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="79788fad-4cc8-442b-9ba6-698e1a05a472" containerName="neutron-httpd" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.703040 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.703438 4678 scope.go:117] "RemoveContainer" containerID="8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.704078 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1\": container with ID starting with 8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1 not found: ID does not exist" containerID="8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.704159 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1"} err="failed to get container status \"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1\": rpc error: code = NotFound desc = could not find container \"8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1\": container with ID starting with 8748d1a7fcc6448b9a1a33986aa8be62525ce1c15c31228ed0e2f551da15cfe1 not found: ID does not exist" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.704268 4678 scope.go:117] "RemoveContainer" containerID="086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f" Oct 13 13:00:32 crc kubenswrapper[4678]: E1013 13:00:32.705031 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f\": container with ID starting with 086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f not found: ID does not exist" containerID="086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.705106 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f"} err="failed to get container status \"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f\": rpc error: code = NotFound desc = could not find container \"086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f\": container with ID starting with 086b867c74bf2431bfdf8a06b930d0bcf35e4aecf6f82169c0d499092753a25f not found: ID does not exist" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.706412 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.726816 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805424 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af635b84-5ac7-42ee-ac5c-55a2c5e20932-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805476 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805548 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805569 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-scripts\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805703 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-886js\" (UniqueName: \"kubernetes.io/projected/af635b84-5ac7-42ee-ac5c-55a2c5e20932-kube-api-access-886js\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.805824 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908110 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908167 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-scripts\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908190 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-886js\" (UniqueName: \"kubernetes.io/projected/af635b84-5ac7-42ee-ac5c-55a2c5e20932-kube-api-access-886js\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908217 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908265 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af635b84-5ac7-42ee-ac5c-55a2c5e20932-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.908300 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.909023 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af635b84-5ac7-42ee-ac5c-55a2c5e20932-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.912981 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.913425 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-scripts\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.913773 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.928078 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af635b84-5ac7-42ee-ac5c-55a2c5e20932-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:32 crc kubenswrapper[4678]: I1013 13:00:32.931788 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-886js\" (UniqueName: \"kubernetes.io/projected/af635b84-5ac7-42ee-ac5c-55a2c5e20932-kube-api-access-886js\") pod \"cinder-scheduler-0\" (UID: \"af635b84-5ac7-42ee-ac5c-55a2c5e20932\") " pod="openstack/cinder-scheduler-0" Oct 13 13:00:33 crc kubenswrapper[4678]: I1013 13:00:33.026188 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 13 13:00:33 crc kubenswrapper[4678]: I1013 13:00:33.115972 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 13 13:00:33 crc kubenswrapper[4678]: I1013 13:00:33.486420 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 13 13:00:33 crc kubenswrapper[4678]: I1013 13:00:33.645502 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"af635b84-5ac7-42ee-ac5c-55a2c5e20932","Type":"ContainerStarted","Data":"82c8b664d3425cbeac6043949fe6ad8c0fa82adc0cc26a6565f7da527933f029"} Oct 13 13:00:33 crc kubenswrapper[4678]: I1013 13:00:33.921496 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-b4bffbf65-fc8vg" Oct 13 13:00:34 crc kubenswrapper[4678]: I1013 13:00:34.608290 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84ac69b4-d0c9-4972-90c4-8232dd9986b1" path="/var/lib/kubelet/pods/84ac69b4-d0c9-4972-90c4-8232dd9986b1/volumes" Oct 13 13:00:34 crc kubenswrapper[4678]: I1013 13:00:34.657672 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"af635b84-5ac7-42ee-ac5c-55a2c5e20932","Type":"ContainerStarted","Data":"fdc71449066f5edf289b0aa717bee39605e8170bf6fd1164c63dac275c1654ff"} Oct 13 13:00:35 crc kubenswrapper[4678]: I1013 13:00:35.671922 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"af635b84-5ac7-42ee-ac5c-55a2c5e20932","Type":"ContainerStarted","Data":"3c91664522f9f346216012c3507afebd2ced5635da99c8c903d904f3be994b5c"} Oct 13 13:00:35 crc kubenswrapper[4678]: I1013 13:00:35.713361 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.713331893 podStartE2EDuration="3.713331893s" podCreationTimestamp="2025-10-13 13:00:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:35.702421507 +0000 UTC m=+1023.786959401" watchObservedRunningTime="2025-10-13 13:00:35.713331893 +0000 UTC m=+1023.797869817" Oct 13 13:00:35 crc kubenswrapper[4678]: I1013 13:00:35.982985 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 13:00:35 crc kubenswrapper[4678]: I1013 13:00:35.983810 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d7ff444f6-wd4jp" podUID="d2954d85-0d57-4f49-91f6-971b0c9005b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.681460 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.682893 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.688764 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.688869 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ssd62" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.692585 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.699572 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.807188 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.807586 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.807636 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.808011 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxc5r\" (UniqueName: \"kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.910504 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxc5r\" (UniqueName: \"kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.910712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.910783 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.910845 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.911781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.919335 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.919641 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.938470 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxc5r\" (UniqueName: \"kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r\") pod \"openstackclient\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.956286 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.957293 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:37 crc kubenswrapper[4678]: I1013 13:00:37.975770 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.027231 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.036185 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.037458 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.047385 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:38 crc kubenswrapper[4678]: E1013 13:00:38.112521 4678 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 13 13:00:38 crc kubenswrapper[4678]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_70008789-6180-47cd-b689-526433f3d780_0(9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a" Netns:"/var/run/netns/e6119854-e109-4ca3-a092-ad30007cb628" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a;K8S_POD_UID=70008789-6180-47cd-b689-526433f3d780" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/70008789-6180-47cd-b689-526433f3d780]: expected pod UID "70008789-6180-47cd-b689-526433f3d780" but got "0b10ccdc-9b5e-4093-8a72-5541ad94e2f3" from Kube API Oct 13 13:00:38 crc kubenswrapper[4678]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 13 13:00:38 crc kubenswrapper[4678]: > Oct 13 13:00:38 crc kubenswrapper[4678]: E1013 13:00:38.112582 4678 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 13 13:00:38 crc kubenswrapper[4678]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_70008789-6180-47cd-b689-526433f3d780_0(9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a" Netns:"/var/run/netns/e6119854-e109-4ca3-a092-ad30007cb628" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=9722fa150b3608b23b8e29c388859ba6215a6ce9b3f76e4f2b3e68a683bfff5a;K8S_POD_UID=70008789-6180-47cd-b689-526433f3d780" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/70008789-6180-47cd-b689-526433f3d780]: expected pod UID "70008789-6180-47cd-b689-526433f3d780" but got "0b10ccdc-9b5e-4093-8a72-5541ad94e2f3" from Kube API Oct 13 13:00:38 crc kubenswrapper[4678]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 13 13:00:38 crc kubenswrapper[4678]: > pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.123819 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.123941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.123979 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8rt\" (UniqueName: \"kubernetes.io/projected/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-kube-api-access-wj8rt\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.124027 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.226374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.226458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8rt\" (UniqueName: \"kubernetes.io/projected/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-kube-api-access-wj8rt\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.226507 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.226573 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.227888 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.230226 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-openstack-config-secret\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.230472 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.243607 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8rt\" (UniqueName: \"kubernetes.io/projected/0b10ccdc-9b5e-4093-8a72-5541ad94e2f3-kube-api-access-wj8rt\") pod \"openstackclient\" (UID: \"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3\") " pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.388922 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.724382 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.740125 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.742964 4678 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="70008789-6180-47cd-b689-526433f3d780" podUID="0b10ccdc-9b5e-4093-8a72-5541ad94e2f3" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.836636 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxc5r\" (UniqueName: \"kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r\") pod \"70008789-6180-47cd-b689-526433f3d780\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.836765 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret\") pod \"70008789-6180-47cd-b689-526433f3d780\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.836842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle\") pod \"70008789-6180-47cd-b689-526433f3d780\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.836887 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config\") pod \"70008789-6180-47cd-b689-526433f3d780\" (UID: \"70008789-6180-47cd-b689-526433f3d780\") " Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.837500 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "70008789-6180-47cd-b689-526433f3d780" (UID: "70008789-6180-47cd-b689-526433f3d780"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.838067 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/70008789-6180-47cd-b689-526433f3d780-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.842107 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "70008789-6180-47cd-b689-526433f3d780" (UID: "70008789-6180-47cd-b689-526433f3d780"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.842619 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r" (OuterVolumeSpecName: "kube-api-access-hxc5r") pod "70008789-6180-47cd-b689-526433f3d780" (UID: "70008789-6180-47cd-b689-526433f3d780"). InnerVolumeSpecName "kube-api-access-hxc5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.843828 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70008789-6180-47cd-b689-526433f3d780" (UID: "70008789-6180-47cd-b689-526433f3d780"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.875277 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.939565 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxc5r\" (UniqueName: \"kubernetes.io/projected/70008789-6180-47cd-b689-526433f3d780-kube-api-access-hxc5r\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.939607 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:38 crc kubenswrapper[4678]: I1013 13:00:38.939622 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70008789-6180-47cd-b689-526433f3d780-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:39 crc kubenswrapper[4678]: I1013 13:00:39.734306 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3","Type":"ContainerStarted","Data":"5f4e9bcf55c0695bf146dfed043ad26d61a75002cadb5f22e400f55e61580643"} Oct 13 13:00:39 crc kubenswrapper[4678]: I1013 13:00:39.734326 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 13 13:00:39 crc kubenswrapper[4678]: I1013 13:00:39.748965 4678 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="70008789-6180-47cd-b689-526433f3d780" podUID="0b10ccdc-9b5e-4093-8a72-5541ad94e2f3" Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.405265 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.405687 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-central-agent" containerID="cri-o://64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74" gracePeriod=30 Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.406881 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="sg-core" containerID="cri-o://72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3" gracePeriod=30 Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.406995 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="proxy-httpd" containerID="cri-o://8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0" gracePeriod=30 Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.407078 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-notification-agent" containerID="cri-o://951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68" gracePeriod=30 Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.580724 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.168:3000/\": EOF" Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.602146 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70008789-6180-47cd-b689-526433f3d780" path="/var/lib/kubelet/pods/70008789-6180-47cd-b689-526433f3d780/volumes" Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.744395 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerID="72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3" exitCode=2 Oct 13 13:00:40 crc kubenswrapper[4678]: I1013 13:00:40.744443 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerDied","Data":"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3"} Oct 13 13:00:41 crc kubenswrapper[4678]: I1013 13:00:41.753537 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerID="8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0" exitCode=0 Oct 13 13:00:41 crc kubenswrapper[4678]: I1013 13:00:41.753766 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerID="64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74" exitCode=0 Oct 13 13:00:41 crc kubenswrapper[4678]: I1013 13:00:41.753569 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerDied","Data":"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0"} Oct 13 13:00:41 crc kubenswrapper[4678]: I1013 13:00:41.753802 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerDied","Data":"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74"} Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.650522 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704277 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk4x6\" (UniqueName: \"kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704446 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704502 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704558 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704596 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704623 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.704657 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd\") pod \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\" (UID: \"8a94d1ed-f73f-48f6-b25e-08c19f3698fb\") " Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.705711 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.706353 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.709819 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6" (OuterVolumeSpecName: "kube-api-access-hk4x6") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "kube-api-access-hk4x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.711826 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts" (OuterVolumeSpecName: "scripts") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.734203 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.769406 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerID="951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68" exitCode=0 Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.769469 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerDied","Data":"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68"} Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.769502 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a94d1ed-f73f-48f6-b25e-08c19f3698fb","Type":"ContainerDied","Data":"09019e02c7f58eceef4f3a5d9529873c5ef3c8842eff06f2aec6ce7f967996c8"} Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.769513 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.769519 4678 scope.go:117] "RemoveContainer" containerID="8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.790453 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806423 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806451 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806461 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806471 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806479 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.806488 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk4x6\" (UniqueName: \"kubernetes.io/projected/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-kube-api-access-hk4x6\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.811027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data" (OuterVolumeSpecName: "config-data") pod "8a94d1ed-f73f-48f6-b25e-08c19f3698fb" (UID: "8a94d1ed-f73f-48f6-b25e-08c19f3698fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:42 crc kubenswrapper[4678]: I1013 13:00:42.908295 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a94d1ed-f73f-48f6-b25e-08c19f3698fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.115781 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.129748 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.148335 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.160639 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:43 crc kubenswrapper[4678]: E1013 13:00:43.161397 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-notification-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.161510 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-notification-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: E1013 13:00:43.161567 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="sg-core" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.161614 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="sg-core" Oct 13 13:00:43 crc kubenswrapper[4678]: E1013 13:00:43.161673 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-central-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.161732 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-central-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: E1013 13:00:43.161790 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="proxy-httpd" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.161843 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="proxy-httpd" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.162107 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-notification-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.171260 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="ceilometer-central-agent" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.171309 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="sg-core" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.171340 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" containerName="proxy-httpd" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.172946 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.173041 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.175127 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.175380 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.284171 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.315732 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5bw2\" (UniqueName: \"kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316196 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316245 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316309 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.316329 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5bw2\" (UniqueName: \"kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418328 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418360 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418380 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418476 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418496 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.418854 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.419692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.424419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.425200 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.426428 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.442420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5bw2\" (UniqueName: \"kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.449110 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data\") pod \"ceilometer-0\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.495637 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.643824 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.644101 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-log" containerID="cri-o://5a3622e06cf6ea08cad1728893e56ff53651623b66d8e1a4d9834187a0431427" gracePeriod=30 Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.644198 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-httpd" containerID="cri-o://51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932" gracePeriod=30 Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.779988 4678 generic.go:334] "Generic (PLEG): container finished" podID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerID="5a3622e06cf6ea08cad1728893e56ff53651623b66d8e1a4d9834187a0431427" exitCode=143 Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.780034 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerDied","Data":"5a3622e06cf6ea08cad1728893e56ff53651623b66d8e1a4d9834187a0431427"} Oct 13 13:00:43 crc kubenswrapper[4678]: I1013 13:00:43.999665 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ftqv2"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.000673 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.015244 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ftqv2"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.098443 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-nhm8n"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.099734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.106346 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nhm8n"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.130485 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sw9m\" (UniqueName: \"kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m\") pod \"nova-api-db-create-ftqv2\" (UID: \"4bc51676-abd0-46b9-8bde-89b0f8d501ce\") " pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.232136 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sw9m\" (UniqueName: \"kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m\") pod \"nova-api-db-create-ftqv2\" (UID: \"4bc51676-abd0-46b9-8bde-89b0f8d501ce\") " pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.232195 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r4bd\" (UniqueName: \"kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd\") pod \"nova-cell0-db-create-nhm8n\" (UID: \"4aac9187-4105-4d36-b4f0-ad03156a500f\") " pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.248622 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sw9m\" (UniqueName: \"kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m\") pod \"nova-api-db-create-ftqv2\" (UID: \"4bc51676-abd0-46b9-8bde-89b0f8d501ce\") " pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.310993 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xmd6j"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.313068 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.321513 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xmd6j"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.321605 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.333243 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r4bd\" (UniqueName: \"kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd\") pod \"nova-cell0-db-create-nhm8n\" (UID: \"4aac9187-4105-4d36-b4f0-ad03156a500f\") " pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.353740 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r4bd\" (UniqueName: \"kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd\") pod \"nova-cell0-db-create-nhm8n\" (UID: \"4aac9187-4105-4d36-b4f0-ad03156a500f\") " pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.415162 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.434785 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twtqj\" (UniqueName: \"kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj\") pod \"nova-cell1-db-create-xmd6j\" (UID: \"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616\") " pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.499649 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5f9fcfccd7-n448c"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.501449 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.508693 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.508927 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.509091 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.531669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5f9fcfccd7-n448c"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.536234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twtqj\" (UniqueName: \"kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj\") pod \"nova-cell1-db-create-xmd6j\" (UID: \"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616\") " pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.550616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twtqj\" (UniqueName: \"kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj\") pod \"nova-cell1-db-create-xmd6j\" (UID: \"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616\") " pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.626554 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a94d1ed-f73f-48f6-b25e-08c19f3698fb" path="/var/lib/kubelet/pods/8a94d1ed-f73f-48f6-b25e-08c19f3698fb/volumes" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.633161 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637360 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-public-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637484 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-etc-swift\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637565 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-config-data\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637697 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-log-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637796 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-combined-ca-bundle\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637860 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zttv\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-kube-api-access-6zttv\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.637931 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-run-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.638008 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-internal-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.739803 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-log-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.739940 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-combined-ca-bundle\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.739969 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zttv\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-kube-api-access-6zttv\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.740007 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-run-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.740043 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-internal-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.740103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-public-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.740135 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-etc-swift\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.740172 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-config-data\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.743423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-config-data\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.743921 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-run-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.746906 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-internal-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.752215 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-public-tls-certs\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.754250 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d9ab7cf-ca73-4d0f-9338-68536f651533-log-httpd\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.757862 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9ab7cf-ca73-4d0f-9338-68536f651533-combined-ca-bundle\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.758532 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zttv\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-kube-api-access-6zttv\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.763023 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6d9ab7cf-ca73-4d0f-9338-68536f651533-etc-swift\") pod \"swift-proxy-5f9fcfccd7-n448c\" (UID: \"6d9ab7cf-ca73-4d0f-9338-68536f651533\") " pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.829017 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.842764 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.842989 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-log" containerID="cri-o://cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e" gracePeriod=30 Oct 13 13:00:44 crc kubenswrapper[4678]: I1013 13:00:44.843450 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-httpd" containerID="cri-o://0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7" gracePeriod=30 Oct 13 13:00:45 crc kubenswrapper[4678]: I1013 13:00:45.518855 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:45 crc kubenswrapper[4678]: I1013 13:00:45.804119 4678 generic.go:334] "Generic (PLEG): container finished" podID="828c58b7-a5c5-4623-8692-e521e61706d8" containerID="cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e" exitCode=143 Oct 13 13:00:45 crc kubenswrapper[4678]: I1013 13:00:45.804163 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerDied","Data":"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e"} Oct 13 13:00:47 crc kubenswrapper[4678]: I1013 13:00:47.828680 4678 generic.go:334] "Generic (PLEG): container finished" podID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerID="51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932" exitCode=0 Oct 13 13:00:47 crc kubenswrapper[4678]: I1013 13:00:47.828823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerDied","Data":"51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932"} Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.336291 4678 scope.go:117] "RemoveContainer" containerID="72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.444178 4678 scope.go:117] "RemoveContainer" containerID="951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.475394 4678 scope.go:117] "RemoveContainer" containerID="64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.521337 4678 scope.go:117] "RemoveContainer" containerID="8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.522320 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0\": container with ID starting with 8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0 not found: ID does not exist" containerID="8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.522368 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0"} err="failed to get container status \"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0\": rpc error: code = NotFound desc = could not find container \"8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0\": container with ID starting with 8df633e13a90040fd5af80418890b1df78c8bfb63968b09f5cf6460679a8f6c0 not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.522402 4678 scope.go:117] "RemoveContainer" containerID="72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.522914 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3\": container with ID starting with 72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3 not found: ID does not exist" containerID="72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.522957 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3"} err="failed to get container status \"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3\": rpc error: code = NotFound desc = could not find container \"72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3\": container with ID starting with 72e2abd62ed26796acb6e8f3e848a7b1dcd1bc0d65c80756e385974da81a65e3 not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.522987 4678 scope.go:117] "RemoveContainer" containerID="951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.523414 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68\": container with ID starting with 951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68 not found: ID does not exist" containerID="951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.523441 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68"} err="failed to get container status \"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68\": rpc error: code = NotFound desc = could not find container \"951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68\": container with ID starting with 951edf8995d4befc0a728d5c328601acf66024433fa5d674756302e566792e68 not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.523455 4678 scope.go:117] "RemoveContainer" containerID="64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.527026 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74\": container with ID starting with 64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74 not found: ID does not exist" containerID="64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.527079 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74"} err="failed to get container status \"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74\": rpc error: code = NotFound desc = could not find container \"64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74\": container with ID starting with 64b9a86c0be2c2c14976233be7aaf256ea93d2c207ec37ef43a132de92254f74 not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.759937 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827250 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827341 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827392 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827439 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827486 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjltj\" (UniqueName: \"kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827540 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827569 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827593 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data\") pod \"828c58b7-a5c5-4623-8692-e521e61706d8\" (UID: \"828c58b7-a5c5-4623-8692-e521e61706d8\") " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.827857 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs" (OuterVolumeSpecName: "logs") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.828124 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.828183 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.835144 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj" (OuterVolumeSpecName: "kube-api-access-rjltj") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "kube-api-access-rjltj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.835216 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.835811 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts" (OuterVolumeSpecName: "scripts") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.845342 4678 generic.go:334] "Generic (PLEG): container finished" podID="828c58b7-a5c5-4623-8692-e521e61706d8" containerID="0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7" exitCode=0 Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.845446 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerDied","Data":"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7"} Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.845482 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"828c58b7-a5c5-4623-8692-e521e61706d8","Type":"ContainerDied","Data":"106e8ae00b24a86d3e2ed6c37b368a0e355bb5af8b1b28d82d9ec03fab07c0f1"} Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.845503 4678 scope.go:117] "RemoveContainer" containerID="0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.845631 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.858260 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0b10ccdc-9b5e-4093-8a72-5541ad94e2f3","Type":"ContainerStarted","Data":"35c36ed24097024afb422b021a95cdc5aa8966ddca18f81db2027e9252b26b2f"} Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.861891 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.881563 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.342657742 podStartE2EDuration="10.881546845s" podCreationTimestamp="2025-10-13 13:00:38 +0000 UTC" firstStartedPulling="2025-10-13 13:00:38.880762185 +0000 UTC m=+1026.965300079" lastFinishedPulling="2025-10-13 13:00:48.419651298 +0000 UTC m=+1036.504189182" observedRunningTime="2025-10-13 13:00:48.880610531 +0000 UTC m=+1036.965148415" watchObservedRunningTime="2025-10-13 13:00:48.881546845 +0000 UTC m=+1036.966084729" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.881867 4678 scope.go:117] "RemoveContainer" containerID="cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.890775 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.893888 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.895394 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data" (OuterVolumeSpecName: "config-data") pod "828c58b7-a5c5-4623-8692-e521e61706d8" (UID: "828c58b7-a5c5-4623-8692-e521e61706d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.902799 4678 scope.go:117] "RemoveContainer" containerID="0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.903634 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7\": container with ID starting with 0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7 not found: ID does not exist" containerID="0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.903666 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7"} err="failed to get container status \"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7\": rpc error: code = NotFound desc = could not find container \"0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7\": container with ID starting with 0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7 not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.903687 4678 scope.go:117] "RemoveContainer" containerID="cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e" Oct 13 13:00:48 crc kubenswrapper[4678]: E1013 13:00:48.904277 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e\": container with ID starting with cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e not found: ID does not exist" containerID="cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.904295 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e"} err="failed to get container status \"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e\": rpc error: code = NotFound desc = could not find container \"cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e\": container with ID starting with cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e not found: ID does not exist" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931447 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjltj\" (UniqueName: \"kubernetes.io/projected/828c58b7-a5c5-4623-8692-e521e61706d8-kube-api-access-rjltj\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931656 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/828c58b7-a5c5-4623-8692-e521e61706d8-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931743 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931801 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931866 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931927 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.931983 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/828c58b7-a5c5-4623-8692-e521e61706d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.970824 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 13 13:00:48 crc kubenswrapper[4678]: I1013 13:00:48.998270 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.008420 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ftqv2"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033101 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033173 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033201 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033320 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033425 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b94c\" (UniqueName: \"kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033444 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033486 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033501 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle\") pod \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\" (UID: \"77ce37dd-9cda-4d4f-9c9e-9c06dff80213\") " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.033826 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.034489 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.035007 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs" (OuterVolumeSpecName: "logs") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.037577 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts" (OuterVolumeSpecName: "scripts") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.041265 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.090250 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c" (OuterVolumeSpecName: "kube-api-access-2b94c") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "kube-api-access-2b94c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.105181 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5f9fcfccd7-n448c"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.130498 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.135778 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.136076 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.136090 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b94c\" (UniqueName: \"kubernetes.io/projected/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-kube-api-access-2b94c\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.136100 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.136127 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.136162 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.194607 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.196137 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.204458 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.218292 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.237274 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.237300 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.257421 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data" (OuterVolumeSpecName: "config-data") pod "77ce37dd-9cda-4d4f-9c9e-9c06dff80213" (UID: "77ce37dd-9cda-4d4f-9c9e-9c06dff80213"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.298581 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:49 crc kubenswrapper[4678]: E1013 13:00:49.298927 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.298940 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: E1013 13:00:49.298954 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.298959 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: E1013 13:00:49.298973 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.298978 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: E1013 13:00:49.299003 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.299008 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.299199 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.299209 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-httpd" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.299227 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.299244 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" containerName="glance-log" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.300219 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.307855 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.308084 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.308765 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-nhm8n"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.321075 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.333009 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xmd6j"] Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.338981 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ce37dd-9cda-4d4f-9c9e-9c06dff80213-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.441940 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.441976 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442030 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442071 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442229 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442309 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442427 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw55x\" (UniqueName: \"kubernetes.io/projected/0b4628b6-c114-4448-b559-79861b376761-kube-api-access-xw55x\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.442657 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-logs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544294 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-logs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544396 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544439 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544534 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544648 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544695 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544759 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw55x\" (UniqueName: \"kubernetes.io/projected/0b4628b6-c114-4448-b559-79861b376761-kube-api-access-xw55x\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.544850 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-logs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.545295 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.545742 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b4628b6-c114-4448-b559-79861b376761-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.552383 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.553978 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.554673 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.555825 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b4628b6-c114-4448-b559-79861b376761-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.560824 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw55x\" (UniqueName: \"kubernetes.io/projected/0b4628b6-c114-4448-b559-79861b376761-kube-api-access-xw55x\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.573620 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0b4628b6-c114-4448-b559-79861b376761\") " pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.618322 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.872317 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bc51676-abd0-46b9-8bde-89b0f8d501ce" containerID="17cc3b99ef9dc7efbe47e41d176a3fc64b8839ad186c3dbb080c967953deecc0" exitCode=0 Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.872391 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ftqv2" event={"ID":"4bc51676-abd0-46b9-8bde-89b0f8d501ce","Type":"ContainerDied","Data":"17cc3b99ef9dc7efbe47e41d176a3fc64b8839ad186c3dbb080c967953deecc0"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.872826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ftqv2" event={"ID":"4bc51676-abd0-46b9-8bde-89b0f8d501ce","Type":"ContainerStarted","Data":"dd6d9a8320c85544011a712a1167a021f0192bbd6f81a03fd181bfb25e239ad5"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.874927 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"77ce37dd-9cda-4d4f-9c9e-9c06dff80213","Type":"ContainerDied","Data":"fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.874967 4678 scope.go:117] "RemoveContainer" containerID="51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.875034 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.894360 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f9fcfccd7-n448c" event={"ID":"6d9ab7cf-ca73-4d0f-9338-68536f651533","Type":"ContainerStarted","Data":"16729596a862b94af89ff374f6c3f7236dfc8c0e086e207a7a4338ed456efeb3"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.894420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f9fcfccd7-n448c" event={"ID":"6d9ab7cf-ca73-4d0f-9338-68536f651533","Type":"ContainerStarted","Data":"61f77c42846847fa63924e911cfb4c85df23605b0f50698ca39cc6661a0a59e9"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.897593 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerStarted","Data":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.897634 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerStarted","Data":"7236a23015965f85d20d23f738e42d877b9980a652fd373c5af3edcf56fd266e"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.899223 4678 generic.go:334] "Generic (PLEG): container finished" podID="6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" containerID="9834e963784cdd5abdcd409fcc1d1623f73c73ed6a621eb5f991c2f5ec833da6" exitCode=0 Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.899276 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xmd6j" event={"ID":"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616","Type":"ContainerDied","Data":"9834e963784cdd5abdcd409fcc1d1623f73c73ed6a621eb5f991c2f5ec833da6"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.899299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xmd6j" event={"ID":"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616","Type":"ContainerStarted","Data":"95a4a9a6c4b3cdaaa8e3f001c1b9d73f7ad96406c75e81b0676be045740cad4f"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.902492 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhm8n" event={"ID":"4aac9187-4105-4d36-b4f0-ad03156a500f","Type":"ContainerStarted","Data":"5e60550dd91d7361c7c27c077becba979569b6bca27c6b1e3e2b14b7ee09b894"} Oct 13 13:00:49 crc kubenswrapper[4678]: I1013 13:00:49.902521 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhm8n" event={"ID":"4aac9187-4105-4d36-b4f0-ad03156a500f","Type":"ContainerStarted","Data":"776ef8e028b00ded1af86946e95a19ca5b43636dd0bb9e1e554bb32d7c18f193"} Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.032317 4678 scope.go:117] "RemoveContainer" containerID="5a3622e06cf6ea08cad1728893e56ff53651623b66d8e1a4d9834187a0431427" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.045512 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.067624 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.088509 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.090030 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.092635 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.092830 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.109606 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164021 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164121 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-logs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164307 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164432 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164483 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164527 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.164570 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ld6s\" (UniqueName: \"kubernetes.io/projected/1b7bcad6-33d5-425e-aeda-70bf6518c557-kube-api-access-8ld6s\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.260947 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268318 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-logs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268421 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268448 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268474 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268499 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ld6s\" (UniqueName: \"kubernetes.io/projected/1b7bcad6-33d5-425e-aeda-70bf6518c557-kube-api-access-8ld6s\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268544 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.268732 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.269071 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.269205 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-logs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.269509 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1b7bcad6-33d5-425e-aeda-70bf6518c557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.274151 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-config-data\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.276721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-scripts\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.277018 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.277552 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b7bcad6-33d5-425e-aeda-70bf6518c557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.287616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ld6s\" (UniqueName: \"kubernetes.io/projected/1b7bcad6-33d5-425e-aeda-70bf6518c557-kube-api-access-8ld6s\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.332883 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"1b7bcad6-33d5-425e-aeda-70bf6518c557\") " pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.419360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.613804 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77ce37dd-9cda-4d4f-9c9e-9c06dff80213" path="/var/lib/kubelet/pods/77ce37dd-9cda-4d4f-9c9e-9c06dff80213/volumes" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.614774 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="828c58b7-a5c5-4623-8692-e521e61706d8" path="/var/lib/kubelet/pods/828c58b7-a5c5-4623-8692-e521e61706d8/volumes" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.916377 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0b4628b6-c114-4448-b559-79861b376761","Type":"ContainerStarted","Data":"ddedbcdb855ca2081a761d5fe1d29069ab4bbd8b4dbac225a881fef8fe4864bf"} Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.921272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5f9fcfccd7-n448c" event={"ID":"6d9ab7cf-ca73-4d0f-9338-68536f651533","Type":"ContainerStarted","Data":"09cda58a2bb8528aac3aa29bd0b64ec449f9fb439e807689f480e3daf44e2065"} Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.922349 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.922377 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.926644 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerStarted","Data":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.933659 4678 generic.go:334] "Generic (PLEG): container finished" podID="4aac9187-4105-4d36-b4f0-ad03156a500f" containerID="5e60550dd91d7361c7c27c077becba979569b6bca27c6b1e3e2b14b7ee09b894" exitCode=0 Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.933945 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhm8n" event={"ID":"4aac9187-4105-4d36-b4f0-ad03156a500f","Type":"ContainerDied","Data":"5e60550dd91d7361c7c27c077becba979569b6bca27c6b1e3e2b14b7ee09b894"} Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.953745 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5f9fcfccd7-n448c" podStartSLOduration=6.953723848 podStartE2EDuration="6.953723848s" podCreationTimestamp="2025-10-13 13:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:50.947519796 +0000 UTC m=+1039.032057680" watchObservedRunningTime="2025-10-13 13:00:50.953723848 +0000 UTC m=+1039.038261732" Oct 13 13:00:50 crc kubenswrapper[4678]: I1013 13:00:50.998036 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.669443 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.675587 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.675784 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.801592 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twtqj\" (UniqueName: \"kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj\") pod \"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616\" (UID: \"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616\") " Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.801801 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r4bd\" (UniqueName: \"kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd\") pod \"4aac9187-4105-4d36-b4f0-ad03156a500f\" (UID: \"4aac9187-4105-4d36-b4f0-ad03156a500f\") " Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.801894 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sw9m\" (UniqueName: \"kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m\") pod \"4bc51676-abd0-46b9-8bde-89b0f8d501ce\" (UID: \"4bc51676-abd0-46b9-8bde-89b0f8d501ce\") " Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.810243 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj" (OuterVolumeSpecName: "kube-api-access-twtqj") pod "6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" (UID: "6bbe76ec-9cc8-4bf1-bda6-ccf07a521616"). InnerVolumeSpecName "kube-api-access-twtqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.813272 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m" (OuterVolumeSpecName: "kube-api-access-7sw9m") pod "4bc51676-abd0-46b9-8bde-89b0f8d501ce" (UID: "4bc51676-abd0-46b9-8bde-89b0f8d501ce"). InnerVolumeSpecName "kube-api-access-7sw9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.817222 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd" (OuterVolumeSpecName: "kube-api-access-8r4bd") pod "4aac9187-4105-4d36-b4f0-ad03156a500f" (UID: "4aac9187-4105-4d36-b4f0-ad03156a500f"). InnerVolumeSpecName "kube-api-access-8r4bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.903771 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r4bd\" (UniqueName: \"kubernetes.io/projected/4aac9187-4105-4d36-b4f0-ad03156a500f-kube-api-access-8r4bd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.903804 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sw9m\" (UniqueName: \"kubernetes.io/projected/4bc51676-abd0-46b9-8bde-89b0f8d501ce-kube-api-access-7sw9m\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.903813 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twtqj\" (UniqueName: \"kubernetes.io/projected/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616-kube-api-access-twtqj\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.939791 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.967321 4678 generic.go:334] "Generic (PLEG): container finished" podID="51a84090-775e-432c-91e0-f142e46147da" containerID="d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a" exitCode=137 Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.967419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerDied","Data":"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a"} Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.967445 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51a84090-775e-432c-91e0-f142e46147da","Type":"ContainerDied","Data":"334d1eb76b660c6db57a3da342c7cf0f9b99af35cea0b751fe654a9e446ea235"} Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.967462 4678 scope.go:117] "RemoveContainer" containerID="d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.967608 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:51 crc kubenswrapper[4678]: I1013 13:00:51.987384 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerStarted","Data":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.003237 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xmd6j" event={"ID":"6bbe76ec-9cc8-4bf1-bda6-ccf07a521616","Type":"ContainerDied","Data":"95a4a9a6c4b3cdaaa8e3f001c1b9d73f7ad96406c75e81b0676be045740cad4f"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.003272 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a4a9a6c4b3cdaaa8e3f001c1b9d73f7ad96406c75e81b0676be045740cad4f" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.003296 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xmd6j" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.023358 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-nhm8n" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.023394 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-nhm8n" event={"ID":"4aac9187-4105-4d36-b4f0-ad03156a500f","Type":"ContainerDied","Data":"776ef8e028b00ded1af86946e95a19ca5b43636dd0bb9e1e554bb32d7c18f193"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.023427 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="776ef8e028b00ded1af86946e95a19ca5b43636dd0bb9e1e554bb32d7c18f193" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.028251 4678 scope.go:117] "RemoveContainer" containerID="da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.031395 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ftqv2" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.031421 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ftqv2" event={"ID":"4bc51676-abd0-46b9-8bde-89b0f8d501ce","Type":"ContainerDied","Data":"dd6d9a8320c85544011a712a1167a021f0192bbd6f81a03fd181bfb25e239ad5"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.031454 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd6d9a8320c85544011a712a1167a021f0192bbd6f81a03fd181bfb25e239ad5" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.033451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0b4628b6-c114-4448-b559-79861b376761","Type":"ContainerStarted","Data":"b0e933ae56c9a07223043b5db5e5d4c1c4c56c0bf0842d3559dae5a09dc42de3"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.037196 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b7bcad6-33d5-425e-aeda-70bf6518c557","Type":"ContainerStarted","Data":"6e69d147d4ad6ea399ee3c2423ee4a5e823a78c0ca0814982640cd027aababd0"} Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.102179 4678 scope.go:117] "RemoveContainer" containerID="d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.105200 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a\": container with ID starting with d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a not found: ID does not exist" containerID="d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.105241 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a"} err="failed to get container status \"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a\": rpc error: code = NotFound desc = could not find container \"d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a\": container with ID starting with d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a not found: ID does not exist" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.105268 4678 scope.go:117] "RemoveContainer" containerID="da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.106570 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee\": container with ID starting with da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee not found: ID does not exist" containerID="da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.106591 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee"} err="failed to get container status \"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee\": rpc error: code = NotFound desc = could not find container \"da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee\": container with ID starting with da7fa4b3241b9aa084e5366a3e47237f2cd1ef9c05a401686d5a857ac81036ee not found: ID does not exist" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110362 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110499 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110583 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110603 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110623 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110654 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.110711 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgjvh\" (UniqueName: \"kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh\") pod \"51a84090-775e-432c-91e0-f142e46147da\" (UID: \"51a84090-775e-432c-91e0-f142e46147da\") " Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.112624 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.112928 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs" (OuterVolumeSpecName: "logs") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.117120 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts" (OuterVolumeSpecName: "scripts") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.117428 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh" (OuterVolumeSpecName: "kube-api-access-lgjvh") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "kube-api-access-lgjvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.130003 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.153743 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.185688 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data" (OuterVolumeSpecName: "config-data") pod "51a84090-775e-432c-91e0-f142e46147da" (UID: "51a84090-775e-432c-91e0-f142e46147da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213111 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213147 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213156 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213164 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a84090-775e-432c-91e0-f142e46147da-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213174 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a84090-775e-432c-91e0-f142e46147da-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213182 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgjvh\" (UniqueName: \"kubernetes.io/projected/51a84090-775e-432c-91e0-f142e46147da-kube-api-access-lgjvh\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.213192 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51a84090-775e-432c-91e0-f142e46147da-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.333247 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.345478 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.356977 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.357355 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357367 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.357385 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357391 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.357408 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api-log" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357414 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api-log" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.357431 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aac9187-4105-4d36-b4f0-ad03156a500f" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357437 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aac9187-4105-4d36-b4f0-ad03156a500f" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: E1013 13:00:52.357453 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc51676-abd0-46b9-8bde-89b0f8d501ce" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357458 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc51676-abd0-46b9-8bde-89b0f8d501ce" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357607 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api-log" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357631 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a84090-775e-432c-91e0-f142e46147da" containerName="cinder-api" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357639 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357650 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc51676-abd0-46b9-8bde-89b0f8d501ce" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.357660 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aac9187-4105-4d36-b4f0-ad03156a500f" containerName="mariadb-database-create" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.358877 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.364089 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.364254 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.364434 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.370924 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519014 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-scripts\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519065 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sxx7\" (UniqueName: \"kubernetes.io/projected/8e246acd-cb1e-42d4-935b-53b777639732-kube-api-access-8sxx7\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519087 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e246acd-cb1e-42d4-935b-53b777639732-logs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519109 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data-custom\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519144 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e246acd-cb1e-42d4-935b-53b777639732-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519161 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519175 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.519299 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.621873 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-scripts\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.621925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sxx7\" (UniqueName: \"kubernetes.io/projected/8e246acd-cb1e-42d4-935b-53b777639732-kube-api-access-8sxx7\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.621956 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e246acd-cb1e-42d4-935b-53b777639732-logs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.621978 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data-custom\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622012 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e246acd-cb1e-42d4-935b-53b777639732-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622031 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622065 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622082 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622140 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622509 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e246acd-cb1e-42d4-935b-53b777639732-logs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.622999 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a84090-775e-432c-91e0-f142e46147da" path="/var/lib/kubelet/pods/51a84090-775e-432c-91e0-f142e46147da/volumes" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.623711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8e246acd-cb1e-42d4-935b-53b777639732-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.629905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data-custom\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.635865 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.636175 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-scripts\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.636700 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.654732 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-config-data\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.660689 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sxx7\" (UniqueName: \"kubernetes.io/projected/8e246acd-cb1e-42d4-935b-53b777639732-kube-api-access-8sxx7\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.671610 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e246acd-cb1e-42d4-935b-53b777639732-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8e246acd-cb1e-42d4-935b-53b777639732\") " pod="openstack/cinder-api-0" Oct 13 13:00:52 crc kubenswrapper[4678]: I1013 13:00:52.685620 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.046914 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0b4628b6-c114-4448-b559-79861b376761","Type":"ContainerStarted","Data":"9c66013a3af8aa2080f6b22e374296308c7713766f9e56cb1d4c62821dbf24d2"} Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.049469 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b7bcad6-33d5-425e-aeda-70bf6518c557","Type":"ContainerStarted","Data":"fa86c7dcc30896e046d32e2f46d40cfb083dcec797f1403a8c63b76617a6303d"} Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.049511 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1b7bcad6-33d5-425e-aeda-70bf6518c557","Type":"ContainerStarted","Data":"185796afee0d2ae31ef759b318bc2b86ff90468d56db88992877300ad2770fd3"} Oct 13 13:00:53 crc kubenswrapper[4678]: E1013 13:00:53.056780 4678 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/f149bbbad5a57dfda75919ec2f267282ac1c0ffe4b989179f639a695508147d0/diff" to get inode usage: stat /var/lib/containers/storage/overlay/f149bbbad5a57dfda75919ec2f267282ac1c0ffe4b989179f639a695508147d0/diff: no such file or directory, extraDiskErr: Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.098526 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.098509196 podStartE2EDuration="3.098509196s" podCreationTimestamp="2025-10-13 13:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:53.096461259 +0000 UTC m=+1041.180999153" watchObservedRunningTime="2025-10-13 13:00:53.098509196 +0000 UTC m=+1041.183047080" Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.098627 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.098623038 podStartE2EDuration="4.098623038s" podCreationTimestamp="2025-10-13 13:00:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:53.072935341 +0000 UTC m=+1041.157473225" watchObservedRunningTime="2025-10-13 13:00:53.098623038 +0000 UTC m=+1041.183160922" Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.115646 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-9bd58f858-cnpb4" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.115910 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:53 crc kubenswrapper[4678]: I1013 13:00:53.169367 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 13 13:00:53 crc kubenswrapper[4678]: E1013 13:00:53.768160 4678 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/596cccc45a4d18f39ba8fa47f6cdbb34e784ce10ea2f67b365bbfb0d42071fef/diff" to get inode usage: stat /var/lib/containers/storage/overlay/596cccc45a4d18f39ba8fa47f6cdbb34e784ce10ea2f67b365bbfb0d42071fef/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-external-api-0_77ce37dd-9cda-4d4f-9c9e-9c06dff80213/glance-log/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-external-api-0_77ce37dd-9cda-4d4f-9c9e-9c06dff80213/glance-log/0.log: no such file or directory Oct 13 13:00:53 crc kubenswrapper[4678]: E1013 13:00:53.870499 4678 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/2afedb5e6112ff4de97547b365f3e37d3b146e9ea81570fe140872fd3be76de7/diff" to get inode usage: stat /var/lib/containers/storage/overlay/2afedb5e6112ff4de97547b365f3e37d3b146e9ea81570fe140872fd3be76de7/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-internal-api-0_828c58b7-a5c5-4623-8692-e521e61706d8/glance-log/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-internal-api-0_828c58b7-a5c5-4623-8692-e521e61706d8/glance-log/0.log: no such file or directory Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.064766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8e246acd-cb1e-42d4-935b-53b777639732","Type":"ContainerStarted","Data":"91c82193bce5b8ac7ffbc8630e8b22baac21052406e7816668ae8ec0c33729ab"} Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.064810 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8e246acd-cb1e-42d4-935b-53b777639732","Type":"ContainerStarted","Data":"05a4cdb3e157b219cc50382d48e7b3ba0cf541475c520861455524ce42300a30"} Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.073805 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-central-agent" containerID="cri-o://8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" gracePeriod=30 Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.073986 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerStarted","Data":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.074710 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.075198 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="proxy-httpd" containerID="cri-o://49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" gracePeriod=30 Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.075249 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="sg-core" containerID="cri-o://1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" gracePeriod=30 Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.075283 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-notification-agent" containerID="cri-o://9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" gracePeriod=30 Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.100071 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=6.626365879 podStartE2EDuration="11.100037415s" podCreationTimestamp="2025-10-13 13:00:43 +0000 UTC" firstStartedPulling="2025-10-13 13:00:49.01193357 +0000 UTC m=+1037.096471444" lastFinishedPulling="2025-10-13 13:00:53.485605096 +0000 UTC m=+1041.570142980" observedRunningTime="2025-10-13 13:00:54.093032064 +0000 UTC m=+1042.177569938" watchObservedRunningTime="2025-10-13 13:00:54.100037415 +0000 UTC m=+1042.184575299" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.197889 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8b33-account-create-597qt"] Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.199475 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.204907 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8b33-account-create-597qt"] Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.207857 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.259770 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spppt\" (UniqueName: \"kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt\") pod \"nova-api-8b33-account-create-597qt\" (UID: \"16d1d3ee-bad0-4228-801e-9f8e14c9170c\") " pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.361596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spppt\" (UniqueName: \"kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt\") pod \"nova-api-8b33-account-create-597qt\" (UID: \"16d1d3ee-bad0-4228-801e-9f8e14c9170c\") " pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.375400 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-6afa-account-create-sf9g6"] Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.376921 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.379230 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.382642 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spppt\" (UniqueName: \"kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt\") pod \"nova-api-8b33-account-create-597qt\" (UID: \"16d1d3ee-bad0-4228-801e-9f8e14c9170c\") " pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.383621 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6afa-account-create-sf9g6"] Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.465070 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn8l6\" (UniqueName: \"kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6\") pod \"nova-cell0-6afa-account-create-sf9g6\" (UID: \"b4d469c9-e858-4fe1-a158-2828b9c35667\") " pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.566927 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn8l6\" (UniqueName: \"kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6\") pod \"nova-cell0-6afa-account-create-sf9g6\" (UID: \"b4d469c9-e858-4fe1-a158-2828b9c35667\") " pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.586376 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn8l6\" (UniqueName: \"kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6\") pod \"nova-cell0-6afa-account-create-sf9g6\" (UID: \"b4d469c9-e858-4fe1-a158-2828b9c35667\") " pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:54 crc kubenswrapper[4678]: E1013 13:00:54.646044 4678 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/36fa1bc27ac6e553b83dfbc75fe1c9060597011ee7cc8868b86e59a4f36b4cab/diff" to get inode usage: stat /var/lib/containers/storage/overlay/36fa1bc27ac6e553b83dfbc75fe1c9060597011ee7cc8868b86e59a4f36b4cab/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-external-api-0_77ce37dd-9cda-4d4f-9c9e-9c06dff80213/glance-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-external-api-0_77ce37dd-9cda-4d4f-9c9e-9c06dff80213/glance-httpd/0.log: no such file or directory Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.670476 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:54 crc kubenswrapper[4678]: E1013 13:00:54.718312 4678 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b24f634647a25f8122653a930ac42ca5e355cf54d8f65992c94c8a1572957507/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b24f634647a25f8122653a930ac42ca5e355cf54d8f65992c94c8a1572957507/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_glance-default-internal-api-0_828c58b7-a5c5-4623-8692-e521e61706d8/glance-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_glance-default-internal-api-0_828c58b7-a5c5-4623-8692-e521e61706d8/glance-httpd/0.log: no such file or directory Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.743013 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.799036 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.839469 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892153 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892409 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5bw2\" (UniqueName: \"kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892617 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892895 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.892996 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.893139 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle\") pod \"093f37f7-ee3d-432e-97ad-3b8b1182f358\" (UID: \"093f37f7-ee3d-432e-97ad-3b8b1182f358\") " Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.894953 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.896090 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.900220 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2" (OuterVolumeSpecName: "kube-api-access-z5bw2") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "kube-api-access-z5bw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.901735 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts" (OuterVolumeSpecName: "scripts") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.919268 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.992812 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996383 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996409 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996420 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5bw2\" (UniqueName: \"kubernetes.io/projected/093f37f7-ee3d-432e-97ad-3b8b1182f358-kube-api-access-z5bw2\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996431 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996439 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/093f37f7-ee3d-432e-97ad-3b8b1182f358-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:54 crc kubenswrapper[4678]: I1013 13:00:54.996446 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.022486 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data" (OuterVolumeSpecName: "config-data") pod "093f37f7-ee3d-432e-97ad-3b8b1182f358" (UID: "093f37f7-ee3d-432e-97ad-3b8b1182f358"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.084906 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8e246acd-cb1e-42d4-935b-53b777639732","Type":"ContainerStarted","Data":"35ec2fc49a1b00b66d196d8150440986c624c7aa5caaeb81d6f93bbdc95f7f08"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.085039 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087820 4678 generic.go:334] "Generic (PLEG): container finished" podID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" exitCode=0 Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087843 4678 generic.go:334] "Generic (PLEG): container finished" podID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" exitCode=2 Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087855 4678 generic.go:334] "Generic (PLEG): container finished" podID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" exitCode=0 Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087862 4678 generic.go:334] "Generic (PLEG): container finished" podID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" exitCode=0 Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087859 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerDied","Data":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087896 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerDied","Data":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087912 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerDied","Data":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087924 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerDied","Data":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087933 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087948 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"093f37f7-ee3d-432e-97ad-3b8b1182f358","Type":"ContainerDied","Data":"7236a23015965f85d20d23f738e42d877b9980a652fd373c5af3edcf56fd266e"} Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.087924 4678 scope.go:117] "RemoveContainer" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.098761 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093f37f7-ee3d-432e-97ad-3b8b1182f358-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.117707 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.117686072 podStartE2EDuration="3.117686072s" podCreationTimestamp="2025-10-13 13:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:00:55.109665379 +0000 UTC m=+1043.194203273" watchObservedRunningTime="2025-10-13 13:00:55.117686072 +0000 UTC m=+1043.202223956" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.133219 4678 scope.go:117] "RemoveContainer" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.137919 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.150191 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.157080 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8b33-account-create-597qt"] Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.161735 4678 scope.go:117] "RemoveContainer" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171087 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.171476 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-notification-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171488 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-notification-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.171504 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="proxy-httpd" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171510 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="proxy-httpd" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.171531 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="sg-core" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171537 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="sg-core" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.171557 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-central-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171563 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-central-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171731 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="proxy-httpd" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171751 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-central-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171761 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="sg-core" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.171771 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" containerName="ceilometer-notification-agent" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.173376 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.175563 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.176784 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.183281 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.231982 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-6afa-account-create-sf9g6"] Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304347 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304441 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304462 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304553 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304612 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbdlq\" (UniqueName: \"kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.304653 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.346467 4678 scope.go:117] "RemoveContainer" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.369652 4678 scope.go:117] "RemoveContainer" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.370035 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": container with ID starting with 49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502 not found: ID does not exist" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370081 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} err="failed to get container status \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": rpc error: code = NotFound desc = could not find container \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": container with ID starting with 49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370107 4678 scope.go:117] "RemoveContainer" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.370395 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": container with ID starting with 1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16 not found: ID does not exist" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370442 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} err="failed to get container status \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": rpc error: code = NotFound desc = could not find container \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": container with ID starting with 1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370469 4678 scope.go:117] "RemoveContainer" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.370709 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": container with ID starting with 9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac not found: ID does not exist" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370735 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} err="failed to get container status \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": rpc error: code = NotFound desc = could not find container \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": container with ID starting with 9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370749 4678 scope.go:117] "RemoveContainer" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: E1013 13:00:55.370968 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": container with ID starting with 8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643 not found: ID does not exist" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.370994 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} err="failed to get container status \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": rpc error: code = NotFound desc = could not find container \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": container with ID starting with 8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371008 4678 scope.go:117] "RemoveContainer" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371310 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} err="failed to get container status \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": rpc error: code = NotFound desc = could not find container \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": container with ID starting with 49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371355 4678 scope.go:117] "RemoveContainer" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371590 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} err="failed to get container status \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": rpc error: code = NotFound desc = could not find container \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": container with ID starting with 1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371612 4678 scope.go:117] "RemoveContainer" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371786 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} err="failed to get container status \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": rpc error: code = NotFound desc = could not find container \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": container with ID starting with 9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371807 4678 scope.go:117] "RemoveContainer" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371946 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} err="failed to get container status \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": rpc error: code = NotFound desc = could not find container \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": container with ID starting with 8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.371963 4678 scope.go:117] "RemoveContainer" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372144 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} err="failed to get container status \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": rpc error: code = NotFound desc = could not find container \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": container with ID starting with 49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372166 4678 scope.go:117] "RemoveContainer" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372359 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} err="failed to get container status \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": rpc error: code = NotFound desc = could not find container \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": container with ID starting with 1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372381 4678 scope.go:117] "RemoveContainer" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372600 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} err="failed to get container status \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": rpc error: code = NotFound desc = could not find container \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": container with ID starting with 9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372623 4678 scope.go:117] "RemoveContainer" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372802 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} err="failed to get container status \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": rpc error: code = NotFound desc = could not find container \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": container with ID starting with 8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372822 4678 scope.go:117] "RemoveContainer" containerID="49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372977 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502"} err="failed to get container status \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": rpc error: code = NotFound desc = could not find container \"49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502\": container with ID starting with 49774aee43b2948f6bec10fbf50edf7d1b4902c3af5bf5e10139ffb213eb6502 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.372996 4678 scope.go:117] "RemoveContainer" containerID="1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.373176 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16"} err="failed to get container status \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": rpc error: code = NotFound desc = could not find container \"1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16\": container with ID starting with 1b231757cc8beae68cbc2fcc8e582ad1e94c1101f0ea65c947ef7bee9fc11f16 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.373191 4678 scope.go:117] "RemoveContainer" containerID="9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.373350 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac"} err="failed to get container status \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": rpc error: code = NotFound desc = could not find container \"9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac\": container with ID starting with 9767243f9b3e041962bc408c591e665a4dfd40870eb8f60cbb0d325378f50fac not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.373372 4678 scope.go:117] "RemoveContainer" containerID="8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.373532 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643"} err="failed to get container status \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": rpc error: code = NotFound desc = could not find container \"8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643\": container with ID starting with 8611fb2393eee41b44974ab9f6c6a7d78345d26e4fb5f3c8de7c1ba582e6e643 not found: ID does not exist" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.405864 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.405904 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.405921 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.405967 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.406022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbdlq\" (UniqueName: \"kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.406063 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.406123 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.406693 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.406794 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.412824 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.413432 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.413705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.421361 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.424370 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbdlq\" (UniqueName: \"kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq\") pod \"ceilometer-0\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " pod="openstack/ceilometer-0" Oct 13 13:00:55 crc kubenswrapper[4678]: I1013 13:00:55.662553 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.107951 4678 generic.go:334] "Generic (PLEG): container finished" podID="16d1d3ee-bad0-4228-801e-9f8e14c9170c" containerID="c8f9f61de16d10814568f1738a0516ee253d9c78f8ccf604d475171b73fe995b" exitCode=0 Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.108114 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8b33-account-create-597qt" event={"ID":"16d1d3ee-bad0-4228-801e-9f8e14c9170c","Type":"ContainerDied","Data":"c8f9f61de16d10814568f1738a0516ee253d9c78f8ccf604d475171b73fe995b"} Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.108144 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8b33-account-create-597qt" event={"ID":"16d1d3ee-bad0-4228-801e-9f8e14c9170c","Type":"ContainerStarted","Data":"04681260deb3a28b911fa232dc84655d0f4fbd0c1dbbe6dd9bf941e60042e68b"} Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.109076 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.110786 4678 generic.go:334] "Generic (PLEG): container finished" podID="b4d469c9-e858-4fe1-a158-2828b9c35667" containerID="71ffbb6871d897e5e68bd12eb6a0988f656723dafd788f5cadb5d2e21068a734" exitCode=0 Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.110858 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6afa-account-create-sf9g6" event={"ID":"b4d469c9-e858-4fe1-a158-2828b9c35667","Type":"ContainerDied","Data":"71ffbb6871d897e5e68bd12eb6a0988f656723dafd788f5cadb5d2e21068a734"} Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.111130 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6afa-account-create-sf9g6" event={"ID":"b4d469c9-e858-4fe1-a158-2828b9c35667","Type":"ContainerStarted","Data":"8817a146ff1452116afad9dffc891a83bb9696bc3a3021d2d0071d6e31d20d4b"} Oct 13 13:00:56 crc kubenswrapper[4678]: I1013 13:00:56.607238 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093f37f7-ee3d-432e-97ad-3b8b1182f358" path="/var/lib/kubelet/pods/093f37f7-ee3d-432e-97ad-3b8b1182f358/volumes" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.119142 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerStarted","Data":"59ad8a8cedd09ab30782e52eda50d0c43ddf8b436dc7a99265d18c4828004e10"} Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.119205 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerStarted","Data":"e1f04e3b3404e8914c14f1987bbb057fdea93e690afb72a78c196647c713fb43"} Oct 13 13:00:57 crc kubenswrapper[4678]: W1013 13:00:57.542426 4678 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod093f37f7_ee3d_432e_97ad_3b8b1182f358.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod093f37f7_ee3d_432e_97ad_3b8b1182f358.slice: no such file or directory Oct 13 13:00:57 crc kubenswrapper[4678]: W1013 13:00:57.546212 4678 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bc51676_abd0_46b9_8bde_89b0f8d501ce.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bc51676_abd0_46b9_8bde_89b0f8d501ce.slice: no such file or directory Oct 13 13:00:57 crc kubenswrapper[4678]: W1013 13:00:57.546260 4678 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aac9187_4105_4d36_b4f0_ad03156a500f.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aac9187_4105_4d36_b4f0_ad03156a500f.slice: no such file or directory Oct 13 13:00:57 crc kubenswrapper[4678]: W1013 13:00:57.546285 4678 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bbe76ec_9cc8_4bf1_bda6_ccf07a521616.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bbe76ec_9cc8_4bf1_bda6_ccf07a521616.slice: no such file or directory Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.655283 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.688922 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.753975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn8l6\" (UniqueName: \"kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6\") pod \"b4d469c9-e858-4fe1-a158-2828b9c35667\" (UID: \"b4d469c9-e858-4fe1-a158-2828b9c35667\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.754185 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spppt\" (UniqueName: \"kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt\") pod \"16d1d3ee-bad0-4228-801e-9f8e14c9170c\" (UID: \"16d1d3ee-bad0-4228-801e-9f8e14c9170c\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.761206 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6" (OuterVolumeSpecName: "kube-api-access-fn8l6") pod "b4d469c9-e858-4fe1-a158-2828b9c35667" (UID: "b4d469c9-e858-4fe1-a158-2828b9c35667"). InnerVolumeSpecName "kube-api-access-fn8l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.773729 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt" (OuterVolumeSpecName: "kube-api-access-spppt") pod "16d1d3ee-bad0-4228-801e-9f8e14c9170c" (UID: "16d1d3ee-bad0-4228-801e-9f8e14c9170c"). InnerVolumeSpecName "kube-api-access-spppt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:57 crc kubenswrapper[4678]: E1013 13:00:57.806762 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice/crio-cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice/crio-106e8ae00b24a86d3e2ed6c37b368a0e355bb5af8b1b28d82d9ec03fab07c0f1\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4901114_e36f_4b5b_8d84_d64f02997156.slice/crio-conmon-4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a84090_775e_432c_91e0_f142e46147da.slice/crio-334d1eb76b660c6db57a3da342c7cf0f9b99af35cea0b751fe654a9e446ea235\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a84090_775e_432c_91e0_f142e46147da.slice/crio-conmon-d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a84090_775e_432c_91e0_f142e46147da.slice/crio-d4c87ed43c4bdcd30dca05b3da56b19ff59a5b00095dcf0408f248991689726a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4901114_e36f_4b5b_8d84_d64f02997156.slice/crio-4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ce37dd_9cda_4d4f_9c9e_9c06dff80213.slice/crio-conmon-51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ce37dd_9cda_4d4f_9c9e_9c06dff80213.slice/crio-51b3b9e06535f8a878baeb7a6b8284607955f8f55009c06e1acb88ce2172a932.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice/crio-conmon-cdceb36374aad2520583e08831550eb75eb5b769c616365824d8ef1de2ed597e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice/crio-0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a84090_775e_432c_91e0_f142e46147da.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ce37dd_9cda_4d4f_9c9e_9c06dff80213.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod828c58b7_a5c5_4623_8692_e521e61706d8.slice/crio-conmon-0e230a287904c2352849e87ee0aac0b61fd615c09b989ff9d10452fe06cad3b7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ce37dd_9cda_4d4f_9c9e_9c06dff80213.slice/crio-fa01feb355c2b41780d194895eda82418df98ebd2c1c15b5c879af0006cd9690\": RecentStats: unable to find data in memory cache]" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.823922 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.856884 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn8l6\" (UniqueName: \"kubernetes.io/projected/b4d469c9-e858-4fe1-a158-2828b9c35667-kube-api-access-fn8l6\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.856905 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spppt\" (UniqueName: \"kubernetes.io/projected/16d1d3ee-bad0-4228-801e-9f8e14c9170c-kube-api-access-spppt\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.960712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtwkn\" (UniqueName: \"kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.960812 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.960867 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.960887 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.960973 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.961807 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs" (OuterVolumeSpecName: "logs") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.961128 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.961925 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts\") pod \"a4901114-e36f-4b5b-8d84-d64f02997156\" (UID: \"a4901114-e36f-4b5b-8d84-d64f02997156\") " Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.962828 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4901114-e36f-4b5b-8d84-d64f02997156-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.965799 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn" (OuterVolumeSpecName: "kube-api-access-mtwkn") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "kube-api-access-mtwkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.967355 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:57 crc kubenswrapper[4678]: I1013 13:00:57.983389 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data" (OuterVolumeSpecName: "config-data") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.007361 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.007823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts" (OuterVolumeSpecName: "scripts") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.029112 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "a4901114-e36f-4b5b-8d84-d64f02997156" (UID: "a4901114-e36f-4b5b-8d84-d64f02997156"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065104 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065289 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065345 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4901114-e36f-4b5b-8d84-d64f02997156-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065398 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065449 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtwkn\" (UniqueName: \"kubernetes.io/projected/a4901114-e36f-4b5b-8d84-d64f02997156-kube-api-access-mtwkn\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.065517 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4901114-e36f-4b5b-8d84-d64f02997156-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.128516 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-6afa-account-create-sf9g6" event={"ID":"b4d469c9-e858-4fe1-a158-2828b9c35667","Type":"ContainerDied","Data":"8817a146ff1452116afad9dffc891a83bb9696bc3a3021d2d0071d6e31d20d4b"} Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.128555 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8817a146ff1452116afad9dffc891a83bb9696bc3a3021d2d0071d6e31d20d4b" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.128568 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-6afa-account-create-sf9g6" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.130717 4678 generic.go:334] "Generic (PLEG): container finished" podID="a4901114-e36f-4b5b-8d84-d64f02997156" containerID="4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71" exitCode=137 Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.130777 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9bd58f858-cnpb4" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.130809 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerDied","Data":"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71"} Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.130846 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9bd58f858-cnpb4" event={"ID":"a4901114-e36f-4b5b-8d84-d64f02997156","Type":"ContainerDied","Data":"c3a911472cc5c980932dc6d9528f853b779f07354bdeb1346939eb29f6c28736"} Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.130871 4678 scope.go:117] "RemoveContainer" containerID="00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.133728 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerStarted","Data":"d8a4d93d5f40513b57a3425a2cbc97832452732acd3418c85c2ed6776ff8d196"} Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.135737 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8b33-account-create-597qt" event={"ID":"16d1d3ee-bad0-4228-801e-9f8e14c9170c","Type":"ContainerDied","Data":"04681260deb3a28b911fa232dc84655d0f4fbd0c1dbbe6dd9bf941e60042e68b"} Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.135767 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04681260deb3a28b911fa232dc84655d0f4fbd0c1dbbe6dd9bf941e60042e68b" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.135811 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8b33-account-create-597qt" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.186206 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.186780 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-9bd58f858-cnpb4"] Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.282378 4678 scope.go:117] "RemoveContainer" containerID="4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.297026 4678 scope.go:117] "RemoveContainer" containerID="00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1" Oct 13 13:00:58 crc kubenswrapper[4678]: E1013 13:00:58.297398 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1\": container with ID starting with 00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1 not found: ID does not exist" containerID="00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.297454 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1"} err="failed to get container status \"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1\": rpc error: code = NotFound desc = could not find container \"00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1\": container with ID starting with 00b2520533c084327efe8206f3f4850a6d66c03c09182dd3e832446cc25165a1 not found: ID does not exist" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.297490 4678 scope.go:117] "RemoveContainer" containerID="4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71" Oct 13 13:00:58 crc kubenswrapper[4678]: E1013 13:00:58.297791 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71\": container with ID starting with 4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71 not found: ID does not exist" containerID="4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.297822 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71"} err="failed to get container status \"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71\": rpc error: code = NotFound desc = could not find container \"4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71\": container with ID starting with 4a28444762b374c35718cb642d6c59f2f981ddfc123f00b77d2012718e73ff71 not found: ID does not exist" Oct 13 13:00:58 crc kubenswrapper[4678]: I1013 13:00:58.625548 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" path="/var/lib/kubelet/pods/a4901114-e36f-4b5b-8d84-d64f02997156/volumes" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.145586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerStarted","Data":"e5d497e6950f9013206abd6bca851554a40aac2377e13d4067f79d7e6d5f2499"} Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.603841 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7l6s4"] Oct 13 13:00:59 crc kubenswrapper[4678]: E1013 13:00:59.604450 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d1d3ee-bad0-4228-801e-9f8e14c9170c" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604467 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d1d3ee-bad0-4228-801e-9f8e14c9170c" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: E1013 13:00:59.604481 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d469c9-e858-4fe1-a158-2828b9c35667" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604488 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d469c9-e858-4fe1-a158-2828b9c35667" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: E1013 13:00:59.604500 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604506 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" Oct 13 13:00:59 crc kubenswrapper[4678]: E1013 13:00:59.604529 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon-log" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604534 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon-log" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604707 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4d469c9-e858-4fe1-a158-2828b9c35667" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604723 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon-log" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604732 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="16d1d3ee-bad0-4228-801e-9f8e14c9170c" containerName="mariadb-account-create" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.604740 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4901114-e36f-4b5b-8d84-d64f02997156" containerName="horizon" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.605422 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.651495 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.651530 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.651647 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.651727 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.651754 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tqp2w" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.664573 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7l6s4"] Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.688551 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.690924 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.691003 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.691171 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.691215 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mdj9\" (UniqueName: \"kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.697131 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.792644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.792722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.792782 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.792803 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mdj9\" (UniqueName: \"kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.799843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.801506 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.801566 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.807860 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mdj9\" (UniqueName: \"kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9\") pod \"nova-cell0-conductor-db-sync-7l6s4\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.847315 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5f9fcfccd7-n448c" Oct 13 13:00:59 crc kubenswrapper[4678]: I1013 13:00:59.986704 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.154199 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29339341-xm4f5"] Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.155433 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.164145 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.164170 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.183244 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29339341-xm4f5"] Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.201267 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.201329 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.201396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw8nd\" (UniqueName: \"kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.201450 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.308103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw8nd\" (UniqueName: \"kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.308177 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.308260 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.308282 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.312834 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.318645 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.319716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.325700 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw8nd\" (UniqueName: \"kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd\") pod \"keystone-cron-29339341-xm4f5\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.420928 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.420964 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.461136 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.468972 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.488984 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.494598 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7l6s4"] Oct 13 13:01:00 crc kubenswrapper[4678]: W1013 13:01:00.507989 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b413c56_81e8_4b62_aa08_742ceca80ec7.slice/crio-c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db WatchSource:0}: Error finding container c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db: Status 404 returned error can't find the container with id c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db Oct 13 13:01:00 crc kubenswrapper[4678]: I1013 13:01:00.959175 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29339341-xm4f5"] Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.181085 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29339341-xm4f5" event={"ID":"2a462e27-6016-43fd-8989-d97b06e1ea3c","Type":"ContainerStarted","Data":"ea6468b10d94bf5c1bd0d48c7ba089a9102b0fe87c52fa6d04f014442c7ac07e"} Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.182690 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" event={"ID":"2b413c56-81e8-4b62-aa08-742ceca80ec7","Type":"ContainerStarted","Data":"c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db"} Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.187863 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerStarted","Data":"fca1ce026e206660ed9a5b45943d441ed219b07a931ebaf78cf896398169f2f6"} Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.188400 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.188426 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.223875 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.276648063 podStartE2EDuration="6.222555751s" podCreationTimestamp="2025-10-13 13:00:55 +0000 UTC" firstStartedPulling="2025-10-13 13:00:56.123224161 +0000 UTC m=+1044.207762045" lastFinishedPulling="2025-10-13 13:01:00.069131849 +0000 UTC m=+1048.153669733" observedRunningTime="2025-10-13 13:01:01.207146589 +0000 UTC m=+1049.291684483" watchObservedRunningTime="2025-10-13 13:01:01.222555751 +0000 UTC m=+1049.307093645" Oct 13 13:01:01 crc kubenswrapper[4678]: I1013 13:01:01.880324 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.196340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29339341-xm4f5" event={"ID":"2a462e27-6016-43fd-8989-d97b06e1ea3c","Type":"ContainerStarted","Data":"04956bdeb207d33f9c1860d0e0ee377c19c2cde95840b6ddc5bd55efe77b581b"} Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.197465 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.196419 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.197623 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.212589 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29339341-xm4f5" podStartSLOduration=2.212569157 podStartE2EDuration="2.212569157s" podCreationTimestamp="2025-10-13 13:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:02.209795033 +0000 UTC m=+1050.294332917" watchObservedRunningTime="2025-10-13 13:01:02.212569157 +0000 UTC m=+1050.297107041" Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.372746 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 13:01:02 crc kubenswrapper[4678]: I1013 13:01:02.442775 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.206568 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-central-agent" containerID="cri-o://59ad8a8cedd09ab30782e52eda50d0c43ddf8b436dc7a99265d18c4828004e10" gracePeriod=30 Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.208028 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="proxy-httpd" containerID="cri-o://fca1ce026e206660ed9a5b45943d441ed219b07a931ebaf78cf896398169f2f6" gracePeriod=30 Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.208089 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-notification-agent" containerID="cri-o://d8a4d93d5f40513b57a3425a2cbc97832452732acd3418c85c2ed6776ff8d196" gracePeriod=30 Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.208148 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="sg-core" containerID="cri-o://e5d497e6950f9013206abd6bca851554a40aac2377e13d4067f79d7e6d5f2499" gracePeriod=30 Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.262332 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 13:01:03 crc kubenswrapper[4678]: I1013 13:01:03.262520 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.076352 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216152 4678 generic.go:334] "Generic (PLEG): container finished" podID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerID="fca1ce026e206660ed9a5b45943d441ed219b07a931ebaf78cf896398169f2f6" exitCode=0 Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216181 4678 generic.go:334] "Generic (PLEG): container finished" podID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerID="e5d497e6950f9013206abd6bca851554a40aac2377e13d4067f79d7e6d5f2499" exitCode=2 Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216189 4678 generic.go:334] "Generic (PLEG): container finished" podID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerID="d8a4d93d5f40513b57a3425a2cbc97832452732acd3418c85c2ed6776ff8d196" exitCode=0 Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216237 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerDied","Data":"fca1ce026e206660ed9a5b45943d441ed219b07a931ebaf78cf896398169f2f6"} Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216287 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerDied","Data":"e5d497e6950f9013206abd6bca851554a40aac2377e13d4067f79d7e6d5f2499"} Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.216299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerDied","Data":"d8a4d93d5f40513b57a3425a2cbc97832452732acd3418c85c2ed6776ff8d196"} Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.218395 4678 generic.go:334] "Generic (PLEG): container finished" podID="2a462e27-6016-43fd-8989-d97b06e1ea3c" containerID="04956bdeb207d33f9c1860d0e0ee377c19c2cde95840b6ddc5bd55efe77b581b" exitCode=0 Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.218638 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29339341-xm4f5" event={"ID":"2a462e27-6016-43fd-8989-d97b06e1ea3c","Type":"ContainerDied","Data":"04956bdeb207d33f9c1860d0e0ee377c19c2cde95840b6ddc5bd55efe77b581b"} Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.515519 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-df52-account-create-qdxzz"] Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.516859 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.519178 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.523976 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-df52-account-create-qdxzz"] Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.598875 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzxqx\" (UniqueName: \"kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx\") pod \"nova-cell1-df52-account-create-qdxzz\" (UID: \"5ba84a93-4108-4b9b-85c0-e4e0e8219398\") " pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.700709 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzxqx\" (UniqueName: \"kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx\") pod \"nova-cell1-df52-account-create-qdxzz\" (UID: \"5ba84a93-4108-4b9b-85c0-e4e0e8219398\") " pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.733685 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzxqx\" (UniqueName: \"kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx\") pod \"nova-cell1-df52-account-create-qdxzz\" (UID: \"5ba84a93-4108-4b9b-85c0-e4e0e8219398\") " pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:04 crc kubenswrapper[4678]: I1013 13:01:04.838965 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.334463 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-df52-account-create-qdxzz"] Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.516346 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.618040 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys\") pod \"2a462e27-6016-43fd-8989-d97b06e1ea3c\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.618157 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw8nd\" (UniqueName: \"kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd\") pod \"2a462e27-6016-43fd-8989-d97b06e1ea3c\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.618207 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle\") pod \"2a462e27-6016-43fd-8989-d97b06e1ea3c\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.618295 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data\") pod \"2a462e27-6016-43fd-8989-d97b06e1ea3c\" (UID: \"2a462e27-6016-43fd-8989-d97b06e1ea3c\") " Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.624174 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd" (OuterVolumeSpecName: "kube-api-access-sw8nd") pod "2a462e27-6016-43fd-8989-d97b06e1ea3c" (UID: "2a462e27-6016-43fd-8989-d97b06e1ea3c"). InnerVolumeSpecName "kube-api-access-sw8nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.624536 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2a462e27-6016-43fd-8989-d97b06e1ea3c" (UID: "2a462e27-6016-43fd-8989-d97b06e1ea3c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.654108 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a462e27-6016-43fd-8989-d97b06e1ea3c" (UID: "2a462e27-6016-43fd-8989-d97b06e1ea3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.684190 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data" (OuterVolumeSpecName: "config-data") pod "2a462e27-6016-43fd-8989-d97b06e1ea3c" (UID: "2a462e27-6016-43fd-8989-d97b06e1ea3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.720514 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.720546 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.720555 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw8nd\" (UniqueName: \"kubernetes.io/projected/2a462e27-6016-43fd-8989-d97b06e1ea3c-kube-api-access-sw8nd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.720565 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a462e27-6016-43fd-8989-d97b06e1ea3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:05 crc kubenswrapper[4678]: I1013 13:01:05.756709 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.236817 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29339341-xm4f5" event={"ID":"2a462e27-6016-43fd-8989-d97b06e1ea3c","Type":"ContainerDied","Data":"ea6468b10d94bf5c1bd0d48c7ba089a9102b0fe87c52fa6d04f014442c7ac07e"} Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.236850 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea6468b10d94bf5c1bd0d48c7ba089a9102b0fe87c52fa6d04f014442c7ac07e" Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.236851 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29339341-xm4f5" Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.240010 4678 generic.go:334] "Generic (PLEG): container finished" podID="5ba84a93-4108-4b9b-85c0-e4e0e8219398" containerID="6993e5d1e6466cefdfa6ff4770a9fe3bce7b274dc3659f6402018d178abe7121" exitCode=0 Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.240072 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-df52-account-create-qdxzz" event={"ID":"5ba84a93-4108-4b9b-85c0-e4e0e8219398","Type":"ContainerDied","Data":"6993e5d1e6466cefdfa6ff4770a9fe3bce7b274dc3659f6402018d178abe7121"} Oct 13 13:01:06 crc kubenswrapper[4678]: I1013 13:01:06.240127 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-df52-account-create-qdxzz" event={"ID":"5ba84a93-4108-4b9b-85c0-e4e0e8219398","Type":"ContainerStarted","Data":"774618073b6bbf5600bd4ad75c0cec3f4b455dca08fd8b0b9704a564fd0b904d"} Oct 13 13:01:07 crc kubenswrapper[4678]: I1013 13:01:07.250146 4678 generic.go:334] "Generic (PLEG): container finished" podID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerID="59ad8a8cedd09ab30782e52eda50d0c43ddf8b436dc7a99265d18c4828004e10" exitCode=0 Oct 13 13:01:07 crc kubenswrapper[4678]: I1013 13:01:07.250221 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerDied","Data":"59ad8a8cedd09ab30782e52eda50d0c43ddf8b436dc7a99265d18c4828004e10"} Oct 13 13:01:10 crc kubenswrapper[4678]: I1013 13:01:10.747930 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:10 crc kubenswrapper[4678]: I1013 13:01:10.939293 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzxqx\" (UniqueName: \"kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx\") pod \"5ba84a93-4108-4b9b-85c0-e4e0e8219398\" (UID: \"5ba84a93-4108-4b9b-85c0-e4e0e8219398\") " Oct 13 13:01:10 crc kubenswrapper[4678]: I1013 13:01:10.942801 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx" (OuterVolumeSpecName: "kube-api-access-xzxqx") pod "5ba84a93-4108-4b9b-85c0-e4e0e8219398" (UID: "5ba84a93-4108-4b9b-85c0-e4e0e8219398"). InnerVolumeSpecName "kube-api-access-xzxqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:10 crc kubenswrapper[4678]: I1013 13:01:10.970904 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.041720 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzxqx\" (UniqueName: \"kubernetes.io/projected/5ba84a93-4108-4b9b-85c0-e4e0e8219398-kube-api-access-xzxqx\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143528 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143653 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbdlq\" (UniqueName: \"kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143697 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143799 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143915 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.143995 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.144018 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.144030 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts\") pod \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\" (UID: \"71dd57e8-39ca-4c04-9fb4-a017abdc2107\") " Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.144514 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.144938 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.144970 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71dd57e8-39ca-4c04-9fb4-a017abdc2107-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.146714 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts" (OuterVolumeSpecName: "scripts") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.147375 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq" (OuterVolumeSpecName: "kube-api-access-zbdlq") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "kube-api-access-zbdlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.166891 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.212503 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.228612 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data" (OuterVolumeSpecName: "config-data") pod "71dd57e8-39ca-4c04-9fb4-a017abdc2107" (UID: "71dd57e8-39ca-4c04-9fb4-a017abdc2107"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.246469 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.246500 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.246509 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbdlq\" (UniqueName: \"kubernetes.io/projected/71dd57e8-39ca-4c04-9fb4-a017abdc2107-kube-api-access-zbdlq\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.246518 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.246526 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71dd57e8-39ca-4c04-9fb4-a017abdc2107-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.298650 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71dd57e8-39ca-4c04-9fb4-a017abdc2107","Type":"ContainerDied","Data":"e1f04e3b3404e8914c14f1987bbb057fdea93e690afb72a78c196647c713fb43"} Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.298710 4678 scope.go:117] "RemoveContainer" containerID="fca1ce026e206660ed9a5b45943d441ed219b07a931ebaf78cf896398169f2f6" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.298710 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.302230 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-df52-account-create-qdxzz" event={"ID":"5ba84a93-4108-4b9b-85c0-e4e0e8219398","Type":"ContainerDied","Data":"774618073b6bbf5600bd4ad75c0cec3f4b455dca08fd8b0b9704a564fd0b904d"} Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.302272 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="774618073b6bbf5600bd4ad75c0cec3f4b455dca08fd8b0b9704a564fd0b904d" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.302341 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-df52-account-create-qdxzz" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.303975 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" event={"ID":"2b413c56-81e8-4b62-aa08-742ceca80ec7","Type":"ContainerStarted","Data":"249bac6a4adbf8689a245ee9be7da465ae58728ba47e9dccf3f685be9e1ea5d4"} Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.340140 4678 scope.go:117] "RemoveContainer" containerID="e5d497e6950f9013206abd6bca851554a40aac2377e13d4067f79d7e6d5f2499" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.357997 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" podStartSLOduration=2.13286187 podStartE2EDuration="12.357976003s" podCreationTimestamp="2025-10-13 13:00:59 +0000 UTC" firstStartedPulling="2025-10-13 13:01:00.512703521 +0000 UTC m=+1048.597241405" lastFinishedPulling="2025-10-13 13:01:10.737817654 +0000 UTC m=+1058.822355538" observedRunningTime="2025-10-13 13:01:11.3333634 +0000 UTC m=+1059.417901284" watchObservedRunningTime="2025-10-13 13:01:11.357976003 +0000 UTC m=+1059.442513897" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.376015 4678 scope.go:117] "RemoveContainer" containerID="d8a4d93d5f40513b57a3425a2cbc97832452732acd3418c85c2ed6776ff8d196" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.383372 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.394897 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.413978 4678 scope.go:117] "RemoveContainer" containerID="59ad8a8cedd09ab30782e52eda50d0c43ddf8b436dc7a99265d18c4828004e10" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415377 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415808 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-notification-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415825 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-notification-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415843 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba84a93-4108-4b9b-85c0-e4e0e8219398" containerName="mariadb-account-create" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415849 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba84a93-4108-4b9b-85c0-e4e0e8219398" containerName="mariadb-account-create" Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415866 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-central-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415872 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-central-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415892 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="sg-core" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415898 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="sg-core" Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415912 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a462e27-6016-43fd-8989-d97b06e1ea3c" containerName="keystone-cron" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415918 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a462e27-6016-43fd-8989-d97b06e1ea3c" containerName="keystone-cron" Oct 13 13:01:11 crc kubenswrapper[4678]: E1013 13:01:11.415927 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="proxy-httpd" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.415934 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="proxy-httpd" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416126 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ba84a93-4108-4b9b-85c0-e4e0e8219398" containerName="mariadb-account-create" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416136 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="sg-core" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416146 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-central-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416169 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a462e27-6016-43fd-8989-d97b06e1ea3c" containerName="keystone-cron" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416178 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="proxy-httpd" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.416187 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" containerName="ceilometer-notification-agent" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.417714 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.420229 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.421163 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.424978 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.553005 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.553086 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.553166 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.553210 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2mdh\" (UniqueName: \"kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.554007 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.554071 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.554132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.660769 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.660832 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.661158 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.661228 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2mdh\" (UniqueName: \"kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.661415 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.661462 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.661542 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.662203 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.662219 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.667273 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.667477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.667543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.674868 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.681858 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2mdh\" (UniqueName: \"kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh\") pod \"ceilometer-0\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " pod="openstack/ceilometer-0" Oct 13 13:01:11 crc kubenswrapper[4678]: I1013 13:01:11.752440 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:12 crc kubenswrapper[4678]: I1013 13:01:12.251401 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:12 crc kubenswrapper[4678]: W1013 13:01:12.252039 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda333c702_3b71_4cdd_9ac8_f65d8a68be5f.slice/crio-9f504dc8df0d29ecfa73dcef0256330a83636a39a89238cc1a38b78e4531ae10 WatchSource:0}: Error finding container 9f504dc8df0d29ecfa73dcef0256330a83636a39a89238cc1a38b78e4531ae10: Status 404 returned error can't find the container with id 9f504dc8df0d29ecfa73dcef0256330a83636a39a89238cc1a38b78e4531ae10 Oct 13 13:01:12 crc kubenswrapper[4678]: I1013 13:01:12.322087 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerStarted","Data":"9f504dc8df0d29ecfa73dcef0256330a83636a39a89238cc1a38b78e4531ae10"} Oct 13 13:01:12 crc kubenswrapper[4678]: I1013 13:01:12.614159 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71dd57e8-39ca-4c04-9fb4-a017abdc2107" path="/var/lib/kubelet/pods/71dd57e8-39ca-4c04-9fb4-a017abdc2107/volumes" Oct 13 13:01:13 crc kubenswrapper[4678]: I1013 13:01:13.338178 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerStarted","Data":"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb"} Oct 13 13:01:15 crc kubenswrapper[4678]: I1013 13:01:15.192081 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:15 crc kubenswrapper[4678]: I1013 13:01:15.360020 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerStarted","Data":"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa"} Oct 13 13:01:15 crc kubenswrapper[4678]: I1013 13:01:15.360090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerStarted","Data":"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306"} Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.383938 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerStarted","Data":"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4"} Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.384127 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-central-agent" containerID="cri-o://dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb" gracePeriod=30 Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.384373 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.384404 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="proxy-httpd" containerID="cri-o://7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4" gracePeriod=30 Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.384471 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="sg-core" containerID="cri-o://45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa" gracePeriod=30 Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.384513 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-notification-agent" containerID="cri-o://cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306" gracePeriod=30 Oct 13 13:01:17 crc kubenswrapper[4678]: I1013 13:01:17.415706 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.86506893 podStartE2EDuration="6.415687564s" podCreationTimestamp="2025-10-13 13:01:11 +0000 UTC" firstStartedPulling="2025-10-13 13:01:12.255172717 +0000 UTC m=+1060.339710611" lastFinishedPulling="2025-10-13 13:01:16.805791361 +0000 UTC m=+1064.890329245" observedRunningTime="2025-10-13 13:01:17.410456835 +0000 UTC m=+1065.494994719" watchObservedRunningTime="2025-10-13 13:01:17.415687564 +0000 UTC m=+1065.500225448" Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.398632 4678 generic.go:334] "Generic (PLEG): container finished" podID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerID="7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4" exitCode=0 Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.399091 4678 generic.go:334] "Generic (PLEG): container finished" podID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerID="45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa" exitCode=2 Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.399107 4678 generic.go:334] "Generic (PLEG): container finished" podID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerID="cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306" exitCode=0 Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.398731 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerDied","Data":"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4"} Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.399153 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerDied","Data":"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa"} Oct 13 13:01:18 crc kubenswrapper[4678]: I1013 13:01:18.399174 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerDied","Data":"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306"} Oct 13 13:01:20 crc kubenswrapper[4678]: I1013 13:01:20.941689 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2mdh\" (UniqueName: \"kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047460 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047549 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047639 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047718 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047812 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.047906 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.050422 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.051045 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.068349 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts" (OuterVolumeSpecName: "scripts") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.068364 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh" (OuterVolumeSpecName: "kube-api-access-s2mdh") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "kube-api-access-s2mdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.077999 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.142989 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data" (OuterVolumeSpecName: "config-data") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.149079 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.149565 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") pod \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\" (UID: \"a333c702-3b71-4cdd-9ac8-f65d8a68be5f\") " Oct 13 13:01:21 crc kubenswrapper[4678]: W1013 13:01:21.149703 4678 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a333c702-3b71-4cdd-9ac8-f65d8a68be5f/volumes/kubernetes.io~secret/combined-ca-bundle Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.149734 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a333c702-3b71-4cdd-9ac8-f65d8a68be5f" (UID: "a333c702-3b71-4cdd-9ac8-f65d8a68be5f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150075 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2mdh\" (UniqueName: \"kubernetes.io/projected/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-kube-api-access-s2mdh\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150096 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150108 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150116 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150125 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150135 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.150142 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a333c702-3b71-4cdd-9ac8-f65d8a68be5f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.441161 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b413c56-81e8-4b62-aa08-742ceca80ec7" containerID="249bac6a4adbf8689a245ee9be7da465ae58728ba47e9dccf3f685be9e1ea5d4" exitCode=0 Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.441233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" event={"ID":"2b413c56-81e8-4b62-aa08-742ceca80ec7","Type":"ContainerDied","Data":"249bac6a4adbf8689a245ee9be7da465ae58728ba47e9dccf3f685be9e1ea5d4"} Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.445396 4678 generic.go:334] "Generic (PLEG): container finished" podID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerID="dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb" exitCode=0 Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.445457 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerDied","Data":"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb"} Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.445509 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.445530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a333c702-3b71-4cdd-9ac8-f65d8a68be5f","Type":"ContainerDied","Data":"9f504dc8df0d29ecfa73dcef0256330a83636a39a89238cc1a38b78e4531ae10"} Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.445565 4678 scope.go:117] "RemoveContainer" containerID="7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.490724 4678 scope.go:117] "RemoveContainer" containerID="45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.508833 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.518804 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.525860 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.526560 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-central-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526586 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-central-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.526617 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="sg-core" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526625 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="sg-core" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.526648 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="proxy-httpd" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526657 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="proxy-httpd" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.526668 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-notification-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526678 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-notification-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526872 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="proxy-httpd" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526895 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-notification-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526913 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="sg-core" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.526923 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" containerName="ceilometer-central-agent" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.528673 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.538412 4678 scope.go:117] "RemoveContainer" containerID="cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.538839 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.539271 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.549323 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.577037 4678 scope.go:117] "RemoveContainer" containerID="dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.601109 4678 scope.go:117] "RemoveContainer" containerID="7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.601477 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4\": container with ID starting with 7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4 not found: ID does not exist" containerID="7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.601510 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4"} err="failed to get container status \"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4\": rpc error: code = NotFound desc = could not find container \"7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4\": container with ID starting with 7a1c79da555a7576af03dcb091248d3e1292742e94cb9c75a280a9f5a7f1d8a4 not found: ID does not exist" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.601548 4678 scope.go:117] "RemoveContainer" containerID="45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.602142 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa\": container with ID starting with 45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa not found: ID does not exist" containerID="45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.602192 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa"} err="failed to get container status \"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa\": rpc error: code = NotFound desc = could not find container \"45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa\": container with ID starting with 45998e5e7400e000aa2ca573f8dfa58272e07be11214ca29bb7240a9008bc4fa not found: ID does not exist" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.602220 4678 scope.go:117] "RemoveContainer" containerID="cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.603008 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306\": container with ID starting with cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306 not found: ID does not exist" containerID="cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.603030 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306"} err="failed to get container status \"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306\": rpc error: code = NotFound desc = could not find container \"cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306\": container with ID starting with cb2f497b456a1917acb2b5acfda9f7d4a3c1a72d180f38181ef303b0c054a306 not found: ID does not exist" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.603078 4678 scope.go:117] "RemoveContainer" containerID="dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb" Oct 13 13:01:21 crc kubenswrapper[4678]: E1013 13:01:21.603529 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb\": container with ID starting with dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb not found: ID does not exist" containerID="dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.603566 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb"} err="failed to get container status \"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb\": rpc error: code = NotFound desc = could not find container \"dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb\": container with ID starting with dc5e4f875bb718d549a4417e1515b9b07564f839484057549f2d2cb7e076f9eb not found: ID does not exist" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659199 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659249 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w46tv\" (UniqueName: \"kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659351 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659456 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659485 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.659631 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761489 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w46tv\" (UniqueName: \"kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761559 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761614 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761639 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761663 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.761695 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.762243 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.762630 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.775177 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.775305 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.775372 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.775626 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.777401 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w46tv\" (UniqueName: \"kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv\") pod \"ceilometer-0\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " pod="openstack/ceilometer-0" Oct 13 13:01:21 crc kubenswrapper[4678]: I1013 13:01:21.867405 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.328261 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:22 crc kubenswrapper[4678]: W1013 13:01:22.339688 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5abad8e7_e210_4981_ad09_e5e175a5f913.slice/crio-4f80e0640c13ccd2824ee91e216bd3ba439c204af9fccc865d6fffc8f9d32fa5 WatchSource:0}: Error finding container 4f80e0640c13ccd2824ee91e216bd3ba439c204af9fccc865d6fffc8f9d32fa5: Status 404 returned error can't find the container with id 4f80e0640c13ccd2824ee91e216bd3ba439c204af9fccc865d6fffc8f9d32fa5 Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.457208 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerStarted","Data":"4f80e0640c13ccd2824ee91e216bd3ba439c204af9fccc865d6fffc8f9d32fa5"} Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.608782 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a333c702-3b71-4cdd-9ac8-f65d8a68be5f" path="/var/lib/kubelet/pods/a333c702-3b71-4cdd-9ac8-f65d8a68be5f/volumes" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.718198 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.881120 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mdj9\" (UniqueName: \"kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9\") pod \"2b413c56-81e8-4b62-aa08-742ceca80ec7\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.881170 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data\") pod \"2b413c56-81e8-4b62-aa08-742ceca80ec7\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.881218 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle\") pod \"2b413c56-81e8-4b62-aa08-742ceca80ec7\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.881282 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts\") pod \"2b413c56-81e8-4b62-aa08-742ceca80ec7\" (UID: \"2b413c56-81e8-4b62-aa08-742ceca80ec7\") " Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.886596 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9" (OuterVolumeSpecName: "kube-api-access-5mdj9") pod "2b413c56-81e8-4b62-aa08-742ceca80ec7" (UID: "2b413c56-81e8-4b62-aa08-742ceca80ec7"). InnerVolumeSpecName "kube-api-access-5mdj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.886806 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts" (OuterVolumeSpecName: "scripts") pod "2b413c56-81e8-4b62-aa08-742ceca80ec7" (UID: "2b413c56-81e8-4b62-aa08-742ceca80ec7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.914732 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data" (OuterVolumeSpecName: "config-data") pod "2b413c56-81e8-4b62-aa08-742ceca80ec7" (UID: "2b413c56-81e8-4b62-aa08-742ceca80ec7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.932288 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b413c56-81e8-4b62-aa08-742ceca80ec7" (UID: "2b413c56-81e8-4b62-aa08-742ceca80ec7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.983155 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mdj9\" (UniqueName: \"kubernetes.io/projected/2b413c56-81e8-4b62-aa08-742ceca80ec7-kube-api-access-5mdj9\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.983186 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.983195 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:22 crc kubenswrapper[4678]: I1013 13:01:22.983204 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b413c56-81e8-4b62-aa08-742ceca80ec7-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.472583 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.472634 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7l6s4" event={"ID":"2b413c56-81e8-4b62-aa08-742ceca80ec7","Type":"ContainerDied","Data":"c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db"} Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.477587 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0a2556d937487dc545ecb083b348a80f2f08caf2e2cc06b59b34ff2e92c04db" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.494337 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerStarted","Data":"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485"} Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.583401 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 13:01:23 crc kubenswrapper[4678]: E1013 13:01:23.583880 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b413c56-81e8-4b62-aa08-742ceca80ec7" containerName="nova-cell0-conductor-db-sync" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.583899 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b413c56-81e8-4b62-aa08-742ceca80ec7" containerName="nova-cell0-conductor-db-sync" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.584190 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b413c56-81e8-4b62-aa08-742ceca80ec7" containerName="nova-cell0-conductor-db-sync" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.585042 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.589505 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.589721 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tqp2w" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.590358 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.695900 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.696093 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.696159 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pq2f\" (UniqueName: \"kubernetes.io/projected/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-kube-api-access-2pq2f\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.797385 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.797659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pq2f\" (UniqueName: \"kubernetes.io/projected/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-kube-api-access-2pq2f\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.798133 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.802118 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.804171 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.814884 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pq2f\" (UniqueName: \"kubernetes.io/projected/0f4da22c-d9e7-4f23-af49-a6ed4551b4b8-kube-api-access-2pq2f\") pod \"nova-cell0-conductor-0\" (UID: \"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8\") " pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:23 crc kubenswrapper[4678]: I1013 13:01:23.962272 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:24 crc kubenswrapper[4678]: I1013 13:01:24.393522 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 13 13:01:24 crc kubenswrapper[4678]: W1013 13:01:24.394876 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f4da22c_d9e7_4f23_af49_a6ed4551b4b8.slice/crio-5aae5f1a8d1543716f7467bdc6a2a1e44ea7ab8d997f6fc152c00f9fda70e942 WatchSource:0}: Error finding container 5aae5f1a8d1543716f7467bdc6a2a1e44ea7ab8d997f6fc152c00f9fda70e942: Status 404 returned error can't find the container with id 5aae5f1a8d1543716f7467bdc6a2a1e44ea7ab8d997f6fc152c00f9fda70e942 Oct 13 13:01:24 crc kubenswrapper[4678]: I1013 13:01:24.509165 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerStarted","Data":"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c"} Oct 13 13:01:24 crc kubenswrapper[4678]: I1013 13:01:24.509211 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerStarted","Data":"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366"} Oct 13 13:01:24 crc kubenswrapper[4678]: I1013 13:01:24.510486 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8","Type":"ContainerStarted","Data":"5aae5f1a8d1543716f7467bdc6a2a1e44ea7ab8d997f6fc152c00f9fda70e942"} Oct 13 13:01:25 crc kubenswrapper[4678]: I1013 13:01:25.506426 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:01:25 crc kubenswrapper[4678]: I1013 13:01:25.506835 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:01:25 crc kubenswrapper[4678]: I1013 13:01:25.518843 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0f4da22c-d9e7-4f23-af49-a6ed4551b4b8","Type":"ContainerStarted","Data":"908ed9dedc368742b3b61e532122d5e810f164c695279e25b3e2046f33722b11"} Oct 13 13:01:25 crc kubenswrapper[4678]: I1013 13:01:25.519042 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:25 crc kubenswrapper[4678]: I1013 13:01:25.536210 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.536193039 podStartE2EDuration="2.536193039s" podCreationTimestamp="2025-10-13 13:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:25.53272358 +0000 UTC m=+1073.617261484" watchObservedRunningTime="2025-10-13 13:01:25.536193039 +0000 UTC m=+1073.620730933" Oct 13 13:01:27 crc kubenswrapper[4678]: I1013 13:01:27.577770 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerStarted","Data":"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25"} Oct 13 13:01:27 crc kubenswrapper[4678]: I1013 13:01:27.580347 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.001969 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.020549 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=9.041143788 podStartE2EDuration="13.020531972s" podCreationTimestamp="2025-10-13 13:01:21 +0000 UTC" firstStartedPulling="2025-10-13 13:01:22.342267354 +0000 UTC m=+1070.426805238" lastFinishedPulling="2025-10-13 13:01:26.321655508 +0000 UTC m=+1074.406193422" observedRunningTime="2025-10-13 13:01:27.612982102 +0000 UTC m=+1075.697519986" watchObservedRunningTime="2025-10-13 13:01:34.020531972 +0000 UTC m=+1082.105069856" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.473163 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-m7csp"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.474724 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.478130 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.478277 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.489008 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m7csp"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.548843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt4hh\" (UniqueName: \"kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.548967 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.549016 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.549092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.650496 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt4hh\" (UniqueName: \"kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.650614 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.650656 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.650686 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.656706 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.657909 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.658657 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.678235 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt4hh\" (UniqueName: \"kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh\") pod \"nova-cell0-cell-mapping-m7csp\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.690979 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.692604 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.701002 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.704231 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.752627 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.752706 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7kbn\" (UniqueName: \"kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.752754 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.814923 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.824557 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.825983 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.832123 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.839226 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.856115 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.856247 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.856287 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7kbn\" (UniqueName: \"kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.866870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.868679 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.894608 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7kbn\" (UniqueName: \"kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn\") pod \"nova-cell1-novncproxy-0\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.923020 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.971593 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd7bk\" (UniqueName: \"kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.971653 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.971723 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.971746 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.974461 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.992117 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 13:01:34 crc kubenswrapper[4678]: I1013 13:01:34.995146 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.010509 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.011886 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.014153 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.020191 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.039206 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.040789 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.048414 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.068014 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074301 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074345 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074424 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074489 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd7bk\" (UniqueName: \"kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074529 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074557 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074589 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87zxp\" (UniqueName: \"kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074613 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg5gx\" (UniqueName: \"kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074632 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.074655 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.075168 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.081284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.094891 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd7bk\" (UniqueName: \"kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.097777 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data\") pod \"nova-metadata-0\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176527 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87zxp\" (UniqueName: \"kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hwmb\" (UniqueName: \"kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg5gx\" (UniqueName: \"kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176669 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176731 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176780 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176849 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176883 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176910 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.176927 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.179517 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.180673 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.183117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.183117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.186706 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.196351 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87zxp\" (UniqueName: \"kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp\") pod \"nova-scheduler-0\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.197432 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg5gx\" (UniqueName: \"kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx\") pod \"nova-api-0\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283121 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283174 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283207 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283225 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hwmb\" (UniqueName: \"kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.283298 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.284360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.284638 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.284874 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.285259 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.285677 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.285782 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.304235 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hwmb\" (UniqueName: \"kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb\") pod \"dnsmasq-dns-845d6d6f59-g72dq\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.348545 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.360484 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.383890 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.429824 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-m7csp"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.541458 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.636896 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dpn5h"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.638156 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.641036 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.641072 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.648014 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dpn5h"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.667977 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"173b4d26-e8d5-4ee8-975a-3b133c992d6e","Type":"ContainerStarted","Data":"e3ad0256838a96d2528e9858076e0cb2b79e24b0f1c6d905cd4971e568295a80"} Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.669322 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m7csp" event={"ID":"9aa05290-ebc1-4886-b08c-0ac9b8c314f0","Type":"ContainerStarted","Data":"c876da047a2d30a0da7f31251005254396976b6c0860b375c7335a0bac2e1e2f"} Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.693098 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.693175 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.693379 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.693526 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk6ks\" (UniqueName: \"kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: W1013 13:01:35.725346 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod893c2b61_85ae_48d0_9914_b85046508949.slice/crio-90ace4eeecacaa2ecb59f601b2c91ebcc1d98fd148bd31ddb67c7e3660e302b3 WatchSource:0}: Error finding container 90ace4eeecacaa2ecb59f601b2c91ebcc1d98fd148bd31ddb67c7e3660e302b3: Status 404 returned error can't find the container with id 90ace4eeecacaa2ecb59f601b2c91ebcc1d98fd148bd31ddb67c7e3660e302b3 Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.727513 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.799183 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.799238 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.799300 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.799343 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk6ks\" (UniqueName: \"kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.807479 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.808369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.816313 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.821133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk6ks\" (UniqueName: \"kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks\") pod \"nova-cell1-conductor-db-sync-dpn5h\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:35 crc kubenswrapper[4678]: W1013 13:01:35.917719 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0296470_6dd6_482f_a782_4d8a24a674a3.slice/crio-59c0e0ae56fbd6ba633def50bcf0fe527b2353a833cf023d8e4f9795cb9e1fa0 WatchSource:0}: Error finding container 59c0e0ae56fbd6ba633def50bcf0fe527b2353a833cf023d8e4f9795cb9e1fa0: Status 404 returned error can't find the container with id 59c0e0ae56fbd6ba633def50bcf0fe527b2353a833cf023d8e4f9795cb9e1fa0 Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.917827 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.930674 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:35 crc kubenswrapper[4678]: I1013 13:01:35.958976 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.043317 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:01:36 crc kubenswrapper[4678]: W1013 13:01:36.080752 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73fbb203_e0e8_46aa_b47d_646507684992.slice/crio-e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826 WatchSource:0}: Error finding container e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826: Status 404 returned error can't find the container with id e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826 Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.392690 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dpn5h"] Oct 13 13:01:36 crc kubenswrapper[4678]: W1013 13:01:36.397339 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb705be0b_2f5e_4778_8160_4f9863ccbda1.slice/crio-7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051 WatchSource:0}: Error finding container 7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051: Status 404 returned error can't find the container with id 7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051 Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.690680 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m7csp" event={"ID":"9aa05290-ebc1-4886-b08c-0ac9b8c314f0","Type":"ContainerStarted","Data":"e9bdcaa3ec8c11fb225963bb6505e33e00b65e603ae23c015784a9ce8aea1491"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.693510 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerStarted","Data":"b4d0ae2e5cb5832ffaf2a0234daa5e4ca6ddba7ee369f3d73617356827da22de"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.694706 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0296470-6dd6-482f-a782-4d8a24a674a3","Type":"ContainerStarted","Data":"59c0e0ae56fbd6ba633def50bcf0fe527b2353a833cf023d8e4f9795cb9e1fa0"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.696214 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerStarted","Data":"90ace4eeecacaa2ecb59f601b2c91ebcc1d98fd148bd31ddb67c7e3660e302b3"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.710642 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" event={"ID":"b705be0b-2f5e-4778-8160-4f9863ccbda1","Type":"ContainerStarted","Data":"aebb155e8ca19274c5d9798de4133f40c054a7e10fff3363cdf162e4be5149af"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.710688 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" event={"ID":"b705be0b-2f5e-4778-8160-4f9863ccbda1","Type":"ContainerStarted","Data":"7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.717001 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-m7csp" podStartSLOduration=2.716980123 podStartE2EDuration="2.716980123s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:36.705622733 +0000 UTC m=+1084.790160617" watchObservedRunningTime="2025-10-13 13:01:36.716980123 +0000 UTC m=+1084.801517997" Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.731637 4678 generic.go:334] "Generic (PLEG): container finished" podID="73fbb203-e0e8-46aa-b47d-646507684992" containerID="32a1c32c9360782e158ab6e81d9fc444a916a8e3a5e1c97580dd75717a0835ec" exitCode=0 Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.731680 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" event={"ID":"73fbb203-e0e8-46aa-b47d-646507684992","Type":"ContainerDied","Data":"32a1c32c9360782e158ab6e81d9fc444a916a8e3a5e1c97580dd75717a0835ec"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.731706 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" event={"ID":"73fbb203-e0e8-46aa-b47d-646507684992","Type":"ContainerStarted","Data":"e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826"} Oct 13 13:01:36 crc kubenswrapper[4678]: I1013 13:01:36.753200 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" podStartSLOduration=1.75317453 podStartE2EDuration="1.75317453s" podCreationTimestamp="2025-10-13 13:01:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:36.73215121 +0000 UTC m=+1084.816689094" watchObservedRunningTime="2025-10-13 13:01:36.75317453 +0000 UTC m=+1084.837712414" Oct 13 13:01:38 crc kubenswrapper[4678]: I1013 13:01:38.751775 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" event={"ID":"73fbb203-e0e8-46aa-b47d-646507684992","Type":"ContainerStarted","Data":"67cee14a7b9c10957db18a2ced5a33dbc9c9f862cce7b649da2cc5958faeaa00"} Oct 13 13:01:38 crc kubenswrapper[4678]: I1013 13:01:38.752143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:38 crc kubenswrapper[4678]: I1013 13:01:38.780910 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" podStartSLOduration=4.780887961 podStartE2EDuration="4.780887961s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:38.771945346 +0000 UTC m=+1086.856483250" watchObservedRunningTime="2025-10-13 13:01:38.780887961 +0000 UTC m=+1086.865425845" Oct 13 13:01:38 crc kubenswrapper[4678]: I1013 13:01:38.852946 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:01:38 crc kubenswrapper[4678]: I1013 13:01:38.858717 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.763507 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"173b4d26-e8d5-4ee8-975a-3b133c992d6e","Type":"ContainerStarted","Data":"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.763934 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b" gracePeriod=30 Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.769230 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerStarted","Data":"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.769269 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerStarted","Data":"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.772451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0296470-6dd6-482f-a782-4d8a24a674a3","Type":"ContainerStarted","Data":"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.775166 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-log" containerID="cri-o://948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55" gracePeriod=30 Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.775349 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerStarted","Data":"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.775385 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerStarted","Data":"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55"} Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.775414 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-metadata" containerID="cri-o://f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d" gracePeriod=30 Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.786848 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.415589552 podStartE2EDuration="5.786824181s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="2025-10-13 13:01:35.560096352 +0000 UTC m=+1083.644634236" lastFinishedPulling="2025-10-13 13:01:38.931330971 +0000 UTC m=+1087.015868865" observedRunningTime="2025-10-13 13:01:39.781222362 +0000 UTC m=+1087.865760266" watchObservedRunningTime="2025-10-13 13:01:39.786824181 +0000 UTC m=+1087.871362065" Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.824856 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.814245727 podStartE2EDuration="5.82483292s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="2025-10-13 13:01:35.921846343 +0000 UTC m=+1084.006384227" lastFinishedPulling="2025-10-13 13:01:38.932433536 +0000 UTC m=+1087.016971420" observedRunningTime="2025-10-13 13:01:39.801367353 +0000 UTC m=+1087.885905247" watchObservedRunningTime="2025-10-13 13:01:39.82483292 +0000 UTC m=+1087.909370804" Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.827803 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.614445629 podStartE2EDuration="5.827782447s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="2025-10-13 13:01:35.728130364 +0000 UTC m=+1083.812668248" lastFinishedPulling="2025-10-13 13:01:38.941467162 +0000 UTC m=+1087.026005066" observedRunningTime="2025-10-13 13:01:39.819358304 +0000 UTC m=+1087.903896178" watchObservedRunningTime="2025-10-13 13:01:39.827782447 +0000 UTC m=+1087.912320331" Oct 13 13:01:39 crc kubenswrapper[4678]: I1013 13:01:39.836930 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.832257339 podStartE2EDuration="5.836915646s" podCreationTimestamp="2025-10-13 13:01:34 +0000 UTC" firstStartedPulling="2025-10-13 13:01:35.925242841 +0000 UTC m=+1084.009780725" lastFinishedPulling="2025-10-13 13:01:38.929901138 +0000 UTC m=+1087.014439032" observedRunningTime="2025-10-13 13:01:39.835697648 +0000 UTC m=+1087.920235532" watchObservedRunningTime="2025-10-13 13:01:39.836915646 +0000 UTC m=+1087.921453530" Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.068957 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.287193 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.287533 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.361615 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.789974 4678 generic.go:334] "Generic (PLEG): container finished" podID="893c2b61-85ae-48d0-9914-b85046508949" containerID="948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55" exitCode=143 Oct 13 13:01:40 crc kubenswrapper[4678]: I1013 13:01:40.790015 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerDied","Data":"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55"} Oct 13 13:01:42 crc kubenswrapper[4678]: I1013 13:01:42.817279 4678 generic.go:334] "Generic (PLEG): container finished" podID="9aa05290-ebc1-4886-b08c-0ac9b8c314f0" containerID="e9bdcaa3ec8c11fb225963bb6505e33e00b65e603ae23c015784a9ce8aea1491" exitCode=0 Oct 13 13:01:42 crc kubenswrapper[4678]: I1013 13:01:42.817403 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m7csp" event={"ID":"9aa05290-ebc1-4886-b08c-0ac9b8c314f0","Type":"ContainerDied","Data":"e9bdcaa3ec8c11fb225963bb6505e33e00b65e603ae23c015784a9ce8aea1491"} Oct 13 13:01:43 crc kubenswrapper[4678]: I1013 13:01:43.833861 4678 generic.go:334] "Generic (PLEG): container finished" podID="b705be0b-2f5e-4778-8160-4f9863ccbda1" containerID="aebb155e8ca19274c5d9798de4133f40c054a7e10fff3363cdf162e4be5149af" exitCode=0 Oct 13 13:01:43 crc kubenswrapper[4678]: I1013 13:01:43.834254 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" event={"ID":"b705be0b-2f5e-4778-8160-4f9863ccbda1","Type":"ContainerDied","Data":"aebb155e8ca19274c5d9798de4133f40c054a7e10fff3363cdf162e4be5149af"} Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.292294 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.367336 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data\") pod \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.367417 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts\") pod \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.367556 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt4hh\" (UniqueName: \"kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh\") pod \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.368633 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle\") pod \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\" (UID: \"9aa05290-ebc1-4886-b08c-0ac9b8c314f0\") " Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.374257 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh" (OuterVolumeSpecName: "kube-api-access-jt4hh") pod "9aa05290-ebc1-4886-b08c-0ac9b8c314f0" (UID: "9aa05290-ebc1-4886-b08c-0ac9b8c314f0"). InnerVolumeSpecName "kube-api-access-jt4hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.375192 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts" (OuterVolumeSpecName: "scripts") pod "9aa05290-ebc1-4886-b08c-0ac9b8c314f0" (UID: "9aa05290-ebc1-4886-b08c-0ac9b8c314f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.402525 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aa05290-ebc1-4886-b08c-0ac9b8c314f0" (UID: "9aa05290-ebc1-4886-b08c-0ac9b8c314f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.412669 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data" (OuterVolumeSpecName: "config-data") pod "9aa05290-ebc1-4886-b08c-0ac9b8c314f0" (UID: "9aa05290-ebc1-4886-b08c-0ac9b8c314f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.470841 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.471188 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.471200 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt4hh\" (UniqueName: \"kubernetes.io/projected/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-kube-api-access-jt4hh\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.471211 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa05290-ebc1-4886-b08c-0ac9b8c314f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.849381 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-m7csp" event={"ID":"9aa05290-ebc1-4886-b08c-0ac9b8c314f0","Type":"ContainerDied","Data":"c876da047a2d30a0da7f31251005254396976b6c0860b375c7335a0bac2e1e2f"} Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.849434 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c876da047a2d30a0da7f31251005254396976b6c0860b375c7335a0bac2e1e2f" Oct 13 13:01:44 crc kubenswrapper[4678]: I1013 13:01:44.849440 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-m7csp" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.083569 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.083865 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-log" containerID="cri-o://151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" gracePeriod=30 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.084347 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-api" containerID="cri-o://ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" gracePeriod=30 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.099312 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.099508 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b0296470-6dd6-482f-a782-4d8a24a674a3" containerName="nova-scheduler-scheduler" containerID="cri-o://adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0" gracePeriod=30 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.387141 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.407779 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.474810 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.475037 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-chq27" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="dnsmasq-dns" containerID="cri-o://d4a1fc9b80b90962e4a82284d80916ad5913d1d1797e076e6b3d2813002e26f3" gracePeriod=10 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.494361 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data\") pod \"b705be0b-2f5e-4778-8160-4f9863ccbda1\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.494473 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk6ks\" (UniqueName: \"kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks\") pod \"b705be0b-2f5e-4778-8160-4f9863ccbda1\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.494646 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle\") pod \"b705be0b-2f5e-4778-8160-4f9863ccbda1\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.494770 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts\") pod \"b705be0b-2f5e-4778-8160-4f9863ccbda1\" (UID: \"b705be0b-2f5e-4778-8160-4f9863ccbda1\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.502660 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks" (OuterVolumeSpecName: "kube-api-access-qk6ks") pod "b705be0b-2f5e-4778-8160-4f9863ccbda1" (UID: "b705be0b-2f5e-4778-8160-4f9863ccbda1"). InnerVolumeSpecName "kube-api-access-qk6ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.508360 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts" (OuterVolumeSpecName: "scripts") pod "b705be0b-2f5e-4778-8160-4f9863ccbda1" (UID: "b705be0b-2f5e-4778-8160-4f9863ccbda1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.535573 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b705be0b-2f5e-4778-8160-4f9863ccbda1" (UID: "b705be0b-2f5e-4778-8160-4f9863ccbda1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.545701 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data" (OuterVolumeSpecName: "config-data") pod "b705be0b-2f5e-4778-8160-4f9863ccbda1" (UID: "b705be0b-2f5e-4778-8160-4f9863ccbda1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.596506 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.596539 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.596549 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk6ks\" (UniqueName: \"kubernetes.io/projected/b705be0b-2f5e-4778-8160-4f9863ccbda1-kube-api-access-qk6ks\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.596558 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b705be0b-2f5e-4778-8160-4f9863ccbda1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.639287 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.698039 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data\") pod \"56f05628-ded2-4d16-93c8-3fd2ff72b397\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.698108 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle\") pod \"56f05628-ded2-4d16-93c8-3fd2ff72b397\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.698374 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg5gx\" (UniqueName: \"kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx\") pod \"56f05628-ded2-4d16-93c8-3fd2ff72b397\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.698515 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs\") pod \"56f05628-ded2-4d16-93c8-3fd2ff72b397\" (UID: \"56f05628-ded2-4d16-93c8-3fd2ff72b397\") " Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.700247 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs" (OuterVolumeSpecName: "logs") pod "56f05628-ded2-4d16-93c8-3fd2ff72b397" (UID: "56f05628-ded2-4d16-93c8-3fd2ff72b397"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.702459 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx" (OuterVolumeSpecName: "kube-api-access-zg5gx") pod "56f05628-ded2-4d16-93c8-3fd2ff72b397" (UID: "56f05628-ded2-4d16-93c8-3fd2ff72b397"). InnerVolumeSpecName "kube-api-access-zg5gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.725544 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data" (OuterVolumeSpecName: "config-data") pod "56f05628-ded2-4d16-93c8-3fd2ff72b397" (UID: "56f05628-ded2-4d16-93c8-3fd2ff72b397"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.730290 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56f05628-ded2-4d16-93c8-3fd2ff72b397" (UID: "56f05628-ded2-4d16-93c8-3fd2ff72b397"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.800192 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.800515 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f05628-ded2-4d16-93c8-3fd2ff72b397-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.800527 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg5gx\" (UniqueName: \"kubernetes.io/projected/56f05628-ded2-4d16-93c8-3fd2ff72b397-kube-api-access-zg5gx\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.800536 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f05628-ded2-4d16-93c8-3fd2ff72b397-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.862320 4678 generic.go:334] "Generic (PLEG): container finished" podID="df8ebd25-714b-4864-ba75-551c20bf5390" containerID="d4a1fc9b80b90962e4a82284d80916ad5913d1d1797e076e6b3d2813002e26f3" exitCode=0 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.862388 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-chq27" event={"ID":"df8ebd25-714b-4864-ba75-551c20bf5390","Type":"ContainerDied","Data":"d4a1fc9b80b90962e4a82284d80916ad5913d1d1797e076e6b3d2813002e26f3"} Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.864527 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" event={"ID":"b705be0b-2f5e-4778-8160-4f9863ccbda1","Type":"ContainerDied","Data":"7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051"} Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.864582 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e360db6294bb1e3af01e361fdb28c6f552da61da7e8250f2ee63f006e8aa051" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.864673 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dpn5h" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874005 4678 generic.go:334] "Generic (PLEG): container finished" podID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerID="ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" exitCode=0 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874032 4678 generic.go:334] "Generic (PLEG): container finished" podID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerID="151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" exitCode=143 Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874064 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerDied","Data":"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4"} Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874085 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerDied","Data":"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6"} Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874096 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"56f05628-ded2-4d16-93c8-3fd2ff72b397","Type":"ContainerDied","Data":"b4d0ae2e5cb5832ffaf2a0234daa5e4ca6ddba7ee369f3d73617356827da22de"} Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874112 4678 scope.go:117] "RemoveContainer" containerID="ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.874250 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.902046 4678 scope.go:117] "RemoveContainer" containerID="151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.904048 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.922806 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.932726 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.936556 4678 scope.go:117] "RemoveContainer" containerID="ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.936985 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4\": container with ID starting with ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4 not found: ID does not exist" containerID="ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.937015 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4"} err="failed to get container status \"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4\": rpc error: code = NotFound desc = could not find container \"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4\": container with ID starting with ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4 not found: ID does not exist" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.937037 4678 scope.go:117] "RemoveContainer" containerID="151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.938934 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6\": container with ID starting with 151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6 not found: ID does not exist" containerID="151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.938969 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6"} err="failed to get container status \"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6\": rpc error: code = NotFound desc = could not find container \"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6\": container with ID starting with 151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6 not found: ID does not exist" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.938985 4678 scope.go:117] "RemoveContainer" containerID="ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939044 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939438 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="init" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939454 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="init" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939466 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa05290-ebc1-4886-b08c-0ac9b8c314f0" containerName="nova-manage" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939472 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa05290-ebc1-4886-b08c-0ac9b8c314f0" containerName="nova-manage" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939487 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b705be0b-2f5e-4778-8160-4f9863ccbda1" containerName="nova-cell1-conductor-db-sync" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939493 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b705be0b-2f5e-4778-8160-4f9863ccbda1" containerName="nova-cell1-conductor-db-sync" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939510 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-api" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939516 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-api" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939533 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-log" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939540 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-log" Oct 13 13:01:45 crc kubenswrapper[4678]: E1013 13:01:45.939560 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="dnsmasq-dns" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939568 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="dnsmasq-dns" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939728 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" containerName="dnsmasq-dns" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939745 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-api" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939758 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b705be0b-2f5e-4778-8160-4f9863ccbda1" containerName="nova-cell1-conductor-db-sync" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939770 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" containerName="nova-api-log" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.939782 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa05290-ebc1-4886-b08c-0ac9b8c314f0" containerName="nova-manage" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.940450 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.940766 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4"} err="failed to get container status \"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4\": rpc error: code = NotFound desc = could not find container \"ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4\": container with ID starting with ef70f1949bd616ad1064934d3255ae174e61de89fb11b61e42f191c203b51ca4 not found: ID does not exist" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.940812 4678 scope.go:117] "RemoveContainer" containerID="151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.941216 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6"} err="failed to get container status \"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6\": rpc error: code = NotFound desc = could not find container \"151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6\": container with ID starting with 151c1d9a4bcfb462938b6d7fdda1b4147405ff6e0e1c76d0e0bc4522afb59cc6 not found: ID does not exist" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.944132 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.944238 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.945607 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.949490 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 13:01:45 crc kubenswrapper[4678]: I1013 13:01:45.977106 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.004648 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.004731 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.004929 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.005001 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xw26\" (UniqueName: \"kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.005074 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.005623 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config\") pod \"df8ebd25-714b-4864-ba75-551c20bf5390\" (UID: \"df8ebd25-714b-4864-ba75-551c20bf5390\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006004 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006142 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006216 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49xxl\" (UniqueName: \"kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006247 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2kkm\" (UniqueName: \"kubernetes.io/projected/902f91dc-30b7-4350-9d03-9517e3b911ae-kube-api-access-t2kkm\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006303 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.006348 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.010269 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26" (OuterVolumeSpecName: "kube-api-access-7xw26") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "kube-api-access-7xw26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.042402 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.051961 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.055467 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config" (OuterVolumeSpecName: "config") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.065953 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.071496 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.075675 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df8ebd25-714b-4864-ba75-551c20bf5390" (UID: "df8ebd25-714b-4864-ba75-551c20bf5390"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108523 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108618 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49xxl\" (UniqueName: \"kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108647 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2kkm\" (UniqueName: \"kubernetes.io/projected/902f91dc-30b7-4350-9d03-9517e3b911ae-kube-api-access-t2kkm\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108699 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108738 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108780 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108917 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108932 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108945 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108958 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xw26\" (UniqueName: \"kubernetes.io/projected/df8ebd25-714b-4864-ba75-551c20bf5390-kube-api-access-7xw26\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108972 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.108983 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df8ebd25-714b-4864-ba75-551c20bf5390-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.109020 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.112024 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.113133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.114231 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.114293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902f91dc-30b7-4350-9d03-9517e3b911ae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.123715 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49xxl\" (UniqueName: \"kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl\") pod \"nova-api-0\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.124727 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2kkm\" (UniqueName: \"kubernetes.io/projected/902f91dc-30b7-4350-9d03-9517e3b911ae-kube-api-access-t2kkm\") pod \"nova-cell1-conductor-0\" (UID: \"902f91dc-30b7-4350-9d03-9517e3b911ae\") " pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.268868 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.291867 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.617623 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f05628-ded2-4d16-93c8-3fd2ff72b397" path="/var/lib/kubelet/pods/56f05628-ded2-4d16-93c8-3fd2ff72b397/volumes" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.728683 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 13 13:01:46 crc kubenswrapper[4678]: W1013 13:01:46.740412 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod902f91dc_30b7_4350_9d03_9517e3b911ae.slice/crio-c9f83332e5c11b04a5341514a312eb1c74611af7a2b5b6096d5c4f212a9cb303 WatchSource:0}: Error finding container c9f83332e5c11b04a5341514a312eb1c74611af7a2b5b6096d5c4f212a9cb303: Status 404 returned error can't find the container with id c9f83332e5c11b04a5341514a312eb1c74611af7a2b5b6096d5c4f212a9cb303 Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.800629 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:01:46 crc kubenswrapper[4678]: W1013 13:01:46.809332 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40584b7d_1111_4a48_afdb_0b76eb151195.slice/crio-c32c4ea52963b1c17b2ddefb93d45fb60b97ff3452b983d89cc31168b6335536 WatchSource:0}: Error finding container c32c4ea52963b1c17b2ddefb93d45fb60b97ff3452b983d89cc31168b6335536: Status 404 returned error can't find the container with id c32c4ea52963b1c17b2ddefb93d45fb60b97ff3452b983d89cc31168b6335536 Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.863534 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.888885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"902f91dc-30b7-4350-9d03-9517e3b911ae","Type":"ContainerStarted","Data":"c9f83332e5c11b04a5341514a312eb1c74611af7a2b5b6096d5c4f212a9cb303"} Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.890480 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerStarted","Data":"c32c4ea52963b1c17b2ddefb93d45fb60b97ff3452b983d89cc31168b6335536"} Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.893100 4678 generic.go:334] "Generic (PLEG): container finished" podID="b0296470-6dd6-482f-a782-4d8a24a674a3" containerID="adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0" exitCode=0 Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.893146 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0296470-6dd6-482f-a782-4d8a24a674a3","Type":"ContainerDied","Data":"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0"} Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.893164 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0296470-6dd6-482f-a782-4d8a24a674a3","Type":"ContainerDied","Data":"59c0e0ae56fbd6ba633def50bcf0fe527b2353a833cf023d8e4f9795cb9e1fa0"} Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.893180 4678 scope.go:117] "RemoveContainer" containerID="adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.893229 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.901512 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-chq27" event={"ID":"df8ebd25-714b-4864-ba75-551c20bf5390","Type":"ContainerDied","Data":"1fdb1fa1fc721c08a9ede97fa5c447dda82697187e7e9f68b2a27c665318bfc9"} Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.901556 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-chq27" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.930073 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle\") pod \"b0296470-6dd6-482f-a782-4d8a24a674a3\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.930175 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data\") pod \"b0296470-6dd6-482f-a782-4d8a24a674a3\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.930241 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87zxp\" (UniqueName: \"kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp\") pod \"b0296470-6dd6-482f-a782-4d8a24a674a3\" (UID: \"b0296470-6dd6-482f-a782-4d8a24a674a3\") " Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.934166 4678 scope.go:117] "RemoveContainer" containerID="adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0" Oct 13 13:01:46 crc kubenswrapper[4678]: E1013 13:01:46.935080 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0\": container with ID starting with adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0 not found: ID does not exist" containerID="adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.935271 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0"} err="failed to get container status \"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0\": rpc error: code = NotFound desc = could not find container \"adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0\": container with ID starting with adf55df99285e5dfd5cfcb51efbfd422061fd20c8df83212f4eee89b76168cf0 not found: ID does not exist" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.935308 4678 scope.go:117] "RemoveContainer" containerID="d4a1fc9b80b90962e4a82284d80916ad5913d1d1797e076e6b3d2813002e26f3" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.937492 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp" (OuterVolumeSpecName: "kube-api-access-87zxp") pod "b0296470-6dd6-482f-a782-4d8a24a674a3" (UID: "b0296470-6dd6-482f-a782-4d8a24a674a3"). InnerVolumeSpecName "kube-api-access-87zxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.943923 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.958171 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-chq27"] Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.964183 4678 scope.go:117] "RemoveContainer" containerID="0281e5471c7e348c85d88c0c81e48eff7380f30838df2998fe72eb41cc94a0f9" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.967343 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data" (OuterVolumeSpecName: "config-data") pod "b0296470-6dd6-482f-a782-4d8a24a674a3" (UID: "b0296470-6dd6-482f-a782-4d8a24a674a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:46 crc kubenswrapper[4678]: I1013 13:01:46.969535 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0296470-6dd6-482f-a782-4d8a24a674a3" (UID: "b0296470-6dd6-482f-a782-4d8a24a674a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.032157 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.032201 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87zxp\" (UniqueName: \"kubernetes.io/projected/b0296470-6dd6-482f-a782-4d8a24a674a3-kube-api-access-87zxp\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.032212 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0296470-6dd6-482f-a782-4d8a24a674a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.279104 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.286430 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.306825 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:47 crc kubenswrapper[4678]: E1013 13:01:47.307213 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0296470-6dd6-482f-a782-4d8a24a674a3" containerName="nova-scheduler-scheduler" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.307232 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0296470-6dd6-482f-a782-4d8a24a674a3" containerName="nova-scheduler-scheduler" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.307411 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0296470-6dd6-482f-a782-4d8a24a674a3" containerName="nova-scheduler-scheduler" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.308027 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.313250 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.317741 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.437523 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.437599 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.437729 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h48gk\" (UniqueName: \"kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.539152 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.539189 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.539330 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h48gk\" (UniqueName: \"kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.544695 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.553591 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.576507 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h48gk\" (UniqueName: \"kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk\") pod \"nova-scheduler-0\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.635432 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.918683 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"902f91dc-30b7-4350-9d03-9517e3b911ae","Type":"ContainerStarted","Data":"71c29c0fcae0290b1b5c80065190635a7fdd3f474276f16aff9604fa947d5528"} Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.919322 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.921785 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerStarted","Data":"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3"} Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.921812 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerStarted","Data":"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1"} Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.944302 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.94428705 podStartE2EDuration="2.94428705s" podCreationTimestamp="2025-10-13 13:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:47.942632332 +0000 UTC m=+1096.027170256" watchObservedRunningTime="2025-10-13 13:01:47.94428705 +0000 UTC m=+1096.028824934" Oct 13 13:01:47 crc kubenswrapper[4678]: I1013 13:01:47.970223 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.970202183 podStartE2EDuration="2.970202183s" podCreationTimestamp="2025-10-13 13:01:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:47.964206806 +0000 UTC m=+1096.048744750" watchObservedRunningTime="2025-10-13 13:01:47.970202183 +0000 UTC m=+1096.054740077" Oct 13 13:01:48 crc kubenswrapper[4678]: I1013 13:01:48.146867 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:01:48 crc kubenswrapper[4678]: I1013 13:01:48.609252 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0296470-6dd6-482f-a782-4d8a24a674a3" path="/var/lib/kubelet/pods/b0296470-6dd6-482f-a782-4d8a24a674a3/volumes" Oct 13 13:01:48 crc kubenswrapper[4678]: I1013 13:01:48.610044 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df8ebd25-714b-4864-ba75-551c20bf5390" path="/var/lib/kubelet/pods/df8ebd25-714b-4864-ba75-551c20bf5390/volumes" Oct 13 13:01:48 crc kubenswrapper[4678]: I1013 13:01:48.934998 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1be79df5-72d5-4616-88e5-62eda59fc6e8","Type":"ContainerStarted","Data":"6e245d433f359286cf38e289d8601e898386d38680a2e55ba65507c557ca5cc3"} Oct 13 13:01:48 crc kubenswrapper[4678]: I1013 13:01:48.935081 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1be79df5-72d5-4616-88e5-62eda59fc6e8","Type":"ContainerStarted","Data":"0ee5f704bce9c3ec9d466e46430819d3453852a164c211424604c73c54559aa4"} Oct 13 13:01:51 crc kubenswrapper[4678]: I1013 13:01:51.315447 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 13 13:01:51 crc kubenswrapper[4678]: I1013 13:01:51.356312 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.356283501 podStartE2EDuration="4.356283501s" podCreationTimestamp="2025-10-13 13:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:01:48.973515112 +0000 UTC m=+1097.058053016" watchObservedRunningTime="2025-10-13 13:01:51.356283501 +0000 UTC m=+1099.440821415" Oct 13 13:01:51 crc kubenswrapper[4678]: I1013 13:01:51.872786 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 13 13:01:52 crc kubenswrapper[4678]: I1013 13:01:52.636283 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 13:01:55 crc kubenswrapper[4678]: I1013 13:01:55.372536 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:55 crc kubenswrapper[4678]: I1013 13:01:55.373224 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" containerName="kube-state-metrics" containerID="cri-o://dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44" gracePeriod=30 Oct 13 13:01:55 crc kubenswrapper[4678]: I1013 13:01:55.508600 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:01:55 crc kubenswrapper[4678]: I1013 13:01:55.508663 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:01:55 crc kubenswrapper[4678]: I1013 13:01:55.941424 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.022771 4678 generic.go:334] "Generic (PLEG): container finished" podID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" containerID="dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44" exitCode=2 Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.022812 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec","Type":"ContainerDied","Data":"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44"} Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.022836 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec","Type":"ContainerDied","Data":"37d36584c35e4061352d83f0e876c469ca98e2b50ee468d917d4ef8e88d29bef"} Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.022852 4678 scope.go:117] "RemoveContainer" containerID="dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.022973 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.052256 4678 scope.go:117] "RemoveContainer" containerID="dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44" Oct 13 13:01:56 crc kubenswrapper[4678]: E1013 13:01:56.052702 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44\": container with ID starting with dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44 not found: ID does not exist" containerID="dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.052743 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44"} err="failed to get container status \"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44\": rpc error: code = NotFound desc = could not find container \"dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44\": container with ID starting with dd91fadaddd68c8d5d386a7da9c7949417545ec41ec3fe4170f1d919cb3afb44 not found: ID does not exist" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.104436 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg6gw\" (UniqueName: \"kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw\") pod \"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec\" (UID: \"3ac0428d-3902-4f97-9ca3-b8bb81ccfeec\") " Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.113868 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw" (OuterVolumeSpecName: "kube-api-access-tg6gw") pod "3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" (UID: "3ac0428d-3902-4f97-9ca3-b8bb81ccfeec"). InnerVolumeSpecName "kube-api-access-tg6gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.206681 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg6gw\" (UniqueName: \"kubernetes.io/projected/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec-kube-api-access-tg6gw\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.292887 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.292975 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.383756 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.400309 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.410184 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:56 crc kubenswrapper[4678]: E1013 13:01:56.410743 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" containerName="kube-state-metrics" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.410763 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" containerName="kube-state-metrics" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.411108 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" containerName="kube-state-metrics" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.411967 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.414247 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.414451 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.419487 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.513549 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.513603 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sq5w\" (UniqueName: \"kubernetes.io/projected/20a83011-dd00-465e-97cc-0bc354fec144-kube-api-access-4sq5w\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.513673 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.513782 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.603322 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ac0428d-3902-4f97-9ca3-b8bb81ccfeec" path="/var/lib/kubelet/pods/3ac0428d-3902-4f97-9ca3-b8bb81ccfeec/volumes" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.615440 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.615511 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.615725 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.615768 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sq5w\" (UniqueName: \"kubernetes.io/projected/20a83011-dd00-465e-97cc-0bc354fec144-kube-api-access-4sq5w\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.621447 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.622216 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.622373 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/20a83011-dd00-465e-97cc-0bc354fec144-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.639874 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sq5w\" (UniqueName: \"kubernetes.io/projected/20a83011-dd00-465e-97cc-0bc354fec144-kube-api-access-4sq5w\") pod \"kube-state-metrics-0\" (UID: \"20a83011-dd00-465e-97cc-0bc354fec144\") " pod="openstack/kube-state-metrics-0" Oct 13 13:01:56 crc kubenswrapper[4678]: I1013 13:01:56.755273 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.202305 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.339425 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.340083 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="proxy-httpd" containerID="cri-o://120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25" gracePeriod=30 Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.340187 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-notification-agent" containerID="cri-o://4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366" gracePeriod=30 Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.340124 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="sg-core" containerID="cri-o://0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c" gracePeriod=30 Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.340493 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-central-agent" containerID="cri-o://5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485" gracePeriod=30 Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.376255 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.376227 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.636155 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 13 13:01:57 crc kubenswrapper[4678]: I1013 13:01:57.717754 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.055722 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a83011-dd00-465e-97cc-0bc354fec144","Type":"ContainerStarted","Data":"bb06019d8498e7888cf1fb2c2b979a8086f9b21f2fcc95514aeedd8dc162d309"} Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.055764 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"20a83011-dd00-465e-97cc-0bc354fec144","Type":"ContainerStarted","Data":"a032519ae7990f4fc0980a98bc41c3b53204328e848c31fbce86231cea7129d7"} Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.055789 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.061856 4678 generic.go:334] "Generic (PLEG): container finished" podID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerID="120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25" exitCode=0 Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.061954 4678 generic.go:334] "Generic (PLEG): container finished" podID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerID="0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c" exitCode=2 Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.062024 4678 generic.go:334] "Generic (PLEG): container finished" podID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerID="5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485" exitCode=0 Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.062607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerDied","Data":"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25"} Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.062749 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerDied","Data":"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c"} Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.062814 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerDied","Data":"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485"} Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.086269 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.707933022 podStartE2EDuration="2.086248802s" podCreationTimestamp="2025-10-13 13:01:56 +0000 UTC" firstStartedPulling="2025-10-13 13:01:57.208446982 +0000 UTC m=+1105.292984866" lastFinishedPulling="2025-10-13 13:01:57.586762752 +0000 UTC m=+1105.671300646" observedRunningTime="2025-10-13 13:01:58.075895445 +0000 UTC m=+1106.160433319" watchObservedRunningTime="2025-10-13 13:01:58.086248802 +0000 UTC m=+1106.170786686" Oct 13 13:01:58 crc kubenswrapper[4678]: I1013 13:01:58.110079 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.486365 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688140 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688335 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w46tv\" (UniqueName: \"kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688485 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688571 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688688 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688786 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688934 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd\") pod \"5abad8e7-e210-4981-ad09-e5e175a5f913\" (UID: \"5abad8e7-e210-4981-ad09-e5e175a5f913\") " Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.688973 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.689421 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.689736 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.695780 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv" (OuterVolumeSpecName: "kube-api-access-w46tv") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "kube-api-access-w46tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.699790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts" (OuterVolumeSpecName: "scripts") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.726779 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.769286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.791387 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.791425 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.791438 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.791452 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5abad8e7-e210-4981-ad09-e5e175a5f913-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.791466 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w46tv\" (UniqueName: \"kubernetes.io/projected/5abad8e7-e210-4981-ad09-e5e175a5f913-kube-api-access-w46tv\") on node \"crc\" DevicePath \"\"" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.808443 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data" (OuterVolumeSpecName: "config-data") pod "5abad8e7-e210-4981-ad09-e5e175a5f913" (UID: "5abad8e7-e210-4981-ad09-e5e175a5f913"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:01:59 crc kubenswrapper[4678]: I1013 13:01:59.893917 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abad8e7-e210-4981-ad09-e5e175a5f913-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.101313 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.101375 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerDied","Data":"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366"} Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.101478 4678 scope.go:117] "RemoveContainer" containerID="120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.103265 4678 generic.go:334] "Generic (PLEG): container finished" podID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerID="4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366" exitCode=0 Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.103355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5abad8e7-e210-4981-ad09-e5e175a5f913","Type":"ContainerDied","Data":"4f80e0640c13ccd2824ee91e216bd3ba439c204af9fccc865d6fffc8f9d32fa5"} Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.133605 4678 scope.go:117] "RemoveContainer" containerID="0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.147823 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.167232 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.178645 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.178992 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="proxy-httpd" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179010 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="proxy-httpd" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.179032 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-notification-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179039 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-notification-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.179088 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-central-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179096 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-central-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.179110 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="sg-core" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179116 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="sg-core" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179285 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="sg-core" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179297 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="proxy-httpd" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179318 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-central-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.179333 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" containerName="ceilometer-notification-agent" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.180131 4678 scope.go:117] "RemoveContainer" containerID="4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.180962 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.184037 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.184223 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.184346 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.207867 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wmqm\" (UniqueName: \"kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.207922 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.207972 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.207989 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.208017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.208041 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.208081 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.208117 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.218863 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.266308 4678 scope.go:117] "RemoveContainer" containerID="5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.288124 4678 scope.go:117] "RemoveContainer" containerID="120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.288833 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25\": container with ID starting with 120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25 not found: ID does not exist" containerID="120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.288885 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25"} err="failed to get container status \"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25\": rpc error: code = NotFound desc = could not find container \"120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25\": container with ID starting with 120b77fe2036e035be2f1af645ea4e7f396ce850cfe4700b2a8b658e90c0fe25 not found: ID does not exist" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.288917 4678 scope.go:117] "RemoveContainer" containerID="0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.289729 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c\": container with ID starting with 0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c not found: ID does not exist" containerID="0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.289772 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c"} err="failed to get container status \"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c\": rpc error: code = NotFound desc = could not find container \"0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c\": container with ID starting with 0fd8b54491d809ecafe04fde65fab5445f7b0cf8c42be7bde541c1212f41366c not found: ID does not exist" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.289802 4678 scope.go:117] "RemoveContainer" containerID="4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.290246 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366\": container with ID starting with 4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366 not found: ID does not exist" containerID="4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.290265 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366"} err="failed to get container status \"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366\": rpc error: code = NotFound desc = could not find container \"4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366\": container with ID starting with 4cef28946d93157624982330bea957fe54a8e5a28f5c2785f5a4266441f57366 not found: ID does not exist" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.290279 4678 scope.go:117] "RemoveContainer" containerID="5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485" Oct 13 13:02:00 crc kubenswrapper[4678]: E1013 13:02:00.290578 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485\": container with ID starting with 5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485 not found: ID does not exist" containerID="5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.290625 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485"} err="failed to get container status \"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485\": rpc error: code = NotFound desc = could not find container \"5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485\": container with ID starting with 5c641bd1669e7cd41c52ce19d2592e53645baef8b34c8acec79a3573cd880485 not found: ID does not exist" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309463 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309520 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wmqm\" (UniqueName: \"kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309612 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309827 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309854 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.309901 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.311147 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.311184 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.316105 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.316739 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.317270 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.317532 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.318345 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.331147 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wmqm\" (UniqueName: \"kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm\") pod \"ceilometer-0\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.561431 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:00 crc kubenswrapper[4678]: I1013 13:02:00.604798 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5abad8e7-e210-4981-ad09-e5e175a5f913" path="/var/lib/kubelet/pods/5abad8e7-e210-4981-ad09-e5e175a5f913/volumes" Oct 13 13:02:01 crc kubenswrapper[4678]: I1013 13:02:01.011642 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:01 crc kubenswrapper[4678]: W1013 13:02:01.012443 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c5f4aa6_d9cf_48ba_8331_9ecd76c24783.slice/crio-a304327ebde9e219f758a12a3738762c6049dcfdcf450fa92e2db627fe6ccca4 WatchSource:0}: Error finding container a304327ebde9e219f758a12a3738762c6049dcfdcf450fa92e2db627fe6ccca4: Status 404 returned error can't find the container with id a304327ebde9e219f758a12a3738762c6049dcfdcf450fa92e2db627fe6ccca4 Oct 13 13:02:01 crc kubenswrapper[4678]: I1013 13:02:01.114343 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerStarted","Data":"a304327ebde9e219f758a12a3738762c6049dcfdcf450fa92e2db627fe6ccca4"} Oct 13 13:02:03 crc kubenswrapper[4678]: I1013 13:02:03.141043 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerStarted","Data":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} Oct 13 13:02:04 crc kubenswrapper[4678]: I1013 13:02:04.152503 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerStarted","Data":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} Oct 13 13:02:04 crc kubenswrapper[4678]: I1013 13:02:04.152551 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerStarted","Data":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.177271 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerStarted","Data":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.177839 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.215313 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9502280920000001 podStartE2EDuration="6.215291936s" podCreationTimestamp="2025-10-13 13:02:00 +0000 UTC" firstStartedPulling="2025-10-13 13:02:01.01696123 +0000 UTC m=+1109.101499124" lastFinishedPulling="2025-10-13 13:02:05.282025044 +0000 UTC m=+1113.366562968" observedRunningTime="2025-10-13 13:02:06.210169083 +0000 UTC m=+1114.294706987" watchObservedRunningTime="2025-10-13 13:02:06.215291936 +0000 UTC m=+1114.299829830" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.296677 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.296950 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.298360 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.298520 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.304431 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.312028 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.499921 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.501863 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.523402 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557469 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557529 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557548 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557564 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn7ng\" (UniqueName: \"kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557585 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.557624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658381 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658437 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658456 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658471 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn7ng\" (UniqueName: \"kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658491 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.658526 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.660098 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.660135 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.660754 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.660968 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.661412 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.699939 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn7ng\" (UniqueName: \"kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng\") pod \"dnsmasq-dns-59cf4bdb65-ssjrr\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.772496 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 13 13:02:06 crc kubenswrapper[4678]: I1013 13:02:06.846026 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:07 crc kubenswrapper[4678]: I1013 13:02:07.317854 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.195460 4678 generic.go:334] "Generic (PLEG): container finished" podID="ef9ff948-801a-4028-a003-47fbe5190158" containerID="83ff36b66a33aa4a0d195f0f9a5ec6c09ac3ad476cdeec458b9001ab52477501" exitCode=0 Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.196936 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" event={"ID":"ef9ff948-801a-4028-a003-47fbe5190158","Type":"ContainerDied","Data":"83ff36b66a33aa4a0d195f0f9a5ec6c09ac3ad476cdeec458b9001ab52477501"} Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.196965 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" event={"ID":"ef9ff948-801a-4028-a003-47fbe5190158","Type":"ContainerStarted","Data":"daff84c8ba73af9bf9bebe678cf2d27abbab0c484cc4d47408e1340233f3a5d7"} Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.377267 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.377816 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-central-agent" containerID="cri-o://b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" gracePeriod=30 Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.378201 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="proxy-httpd" containerID="cri-o://bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" gracePeriod=30 Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.378241 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="sg-core" containerID="cri-o://49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" gracePeriod=30 Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.378272 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-notification-agent" containerID="cri-o://61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" gracePeriod=30 Oct 13 13:02:08 crc kubenswrapper[4678]: I1013 13:02:08.931948 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.180689 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.210319 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" event={"ID":"ef9ff948-801a-4028-a003-47fbe5190158","Type":"ContainerStarted","Data":"4daecfe5e5fedc235149421381db7b1c6490e1e7548dbe656f57f2e1338c946e"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.210443 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212748 4678 generic.go:334] "Generic (PLEG): container finished" podID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" exitCode=0 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212772 4678 generic.go:334] "Generic (PLEG): container finished" podID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" exitCode=2 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212782 4678 generic.go:334] "Generic (PLEG): container finished" podID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" exitCode=0 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212789 4678 generic.go:334] "Generic (PLEG): container finished" podID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" exitCode=0 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212794 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212833 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerDied","Data":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerDied","Data":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212897 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerDied","Data":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212908 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerDied","Data":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783","Type":"ContainerDied","Data":"a304327ebde9e219f758a12a3738762c6049dcfdcf450fa92e2db627fe6ccca4"} Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.212935 4678 scope.go:117] "RemoveContainer" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.213097 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-log" containerID="cri-o://2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1" gracePeriod=30 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.213121 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-api" containerID="cri-o://656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3" gracePeriod=30 Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.237390 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" podStartSLOduration=3.237374477 podStartE2EDuration="3.237374477s" podCreationTimestamp="2025-10-13 13:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:09.234459451 +0000 UTC m=+1117.318997335" watchObservedRunningTime="2025-10-13 13:02:09.237374477 +0000 UTC m=+1117.321912361" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.246968 4678 scope.go:117] "RemoveContainer" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.265958 4678 scope.go:117] "RemoveContainer" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.294481 4678 scope.go:117] "RemoveContainer" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.312949 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313180 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wmqm\" (UniqueName: \"kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313235 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313286 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313514 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313558 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313589 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.313682 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml\") pod \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\" (UID: \"3c5f4aa6-d9cf-48ba-8331-9ecd76c24783\") " Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.314722 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.314852 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.316495 4678 scope.go:117] "RemoveContainer" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.318295 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": container with ID starting with bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a not found: ID does not exist" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.318353 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} err="failed to get container status \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": rpc error: code = NotFound desc = could not find container \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": container with ID starting with bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.318384 4678 scope.go:117] "RemoveContainer" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.318822 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": container with ID starting with 49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6 not found: ID does not exist" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.318852 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} err="failed to get container status \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": rpc error: code = NotFound desc = could not find container \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": container with ID starting with 49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.318871 4678 scope.go:117] "RemoveContainer" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.319697 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm" (OuterVolumeSpecName: "kube-api-access-4wmqm") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "kube-api-access-4wmqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.320975 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts" (OuterVolumeSpecName: "scripts") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.322220 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": container with ID starting with 61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434 not found: ID does not exist" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.322262 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} err="failed to get container status \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": rpc error: code = NotFound desc = could not find container \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": container with ID starting with 61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.322287 4678 scope.go:117] "RemoveContainer" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.322845 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": container with ID starting with b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42 not found: ID does not exist" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.322871 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} err="failed to get container status \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": rpc error: code = NotFound desc = could not find container \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": container with ID starting with b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.322889 4678 scope.go:117] "RemoveContainer" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.323164 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} err="failed to get container status \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": rpc error: code = NotFound desc = could not find container \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": container with ID starting with bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.323185 4678 scope.go:117] "RemoveContainer" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.324407 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} err="failed to get container status \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": rpc error: code = NotFound desc = could not find container \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": container with ID starting with 49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.324443 4678 scope.go:117] "RemoveContainer" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.324720 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} err="failed to get container status \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": rpc error: code = NotFound desc = could not find container \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": container with ID starting with 61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.324749 4678 scope.go:117] "RemoveContainer" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.325610 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} err="failed to get container status \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": rpc error: code = NotFound desc = could not find container \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": container with ID starting with b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.325649 4678 scope.go:117] "RemoveContainer" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.326152 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} err="failed to get container status \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": rpc error: code = NotFound desc = could not find container \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": container with ID starting with bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.326171 4678 scope.go:117] "RemoveContainer" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.326436 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} err="failed to get container status \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": rpc error: code = NotFound desc = could not find container \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": container with ID starting with 49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.326455 4678 scope.go:117] "RemoveContainer" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.328013 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} err="failed to get container status \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": rpc error: code = NotFound desc = could not find container \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": container with ID starting with 61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.328030 4678 scope.go:117] "RemoveContainer" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.329025 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} err="failed to get container status \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": rpc error: code = NotFound desc = could not find container \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": container with ID starting with b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.329039 4678 scope.go:117] "RemoveContainer" containerID="bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.330138 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a"} err="failed to get container status \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": rpc error: code = NotFound desc = could not find container \"bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a\": container with ID starting with bdead9132702d16e1ddbc7b5e3315ff83f0e31d024ae06a081b0f6fef3e8ac4a not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.330160 4678 scope.go:117] "RemoveContainer" containerID="49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.331441 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6"} err="failed to get container status \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": rpc error: code = NotFound desc = could not find container \"49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6\": container with ID starting with 49d0afe9186d24d944d14afc05c4eac09ca3d7e84d5ce2dae3251d1cc4803cf6 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.331461 4678 scope.go:117] "RemoveContainer" containerID="61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.331920 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434"} err="failed to get container status \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": rpc error: code = NotFound desc = could not find container \"61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434\": container with ID starting with 61e1fba61e15ff471128d352c284719bf7c2bd719f162d7f1e1a8820b3dbd434 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.331948 4678 scope.go:117] "RemoveContainer" containerID="b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.332154 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42"} err="failed to get container status \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": rpc error: code = NotFound desc = could not find container \"b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42\": container with ID starting with b2660d68ab7ef0426dd01701fb51870ad7e0d1dfd294b3a6ca81cef67a302a42 not found: ID does not exist" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.350990 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.372741 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.402009 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.416889 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.417164 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.417287 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wmqm\" (UniqueName: \"kubernetes.io/projected/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-kube-api-access-4wmqm\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.417382 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.421436 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.422593 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.422737 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.431246 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data" (OuterVolumeSpecName: "config-data") pod "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" (UID: "3c5f4aa6-d9cf-48ba-8331-9ecd76c24783"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.528888 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.555902 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.570239 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.580175 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.580751 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="sg-core" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.580768 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="sg-core" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.582302 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-central-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.582345 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-central-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.582394 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-notification-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.582404 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-notification-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.582433 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="proxy-httpd" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.582442 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="proxy-httpd" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.582950 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="proxy-httpd" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.582981 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-notification-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.583010 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="ceilometer-central-agent" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.583038 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" containerName="sg-core" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.616612 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.620987 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.621200 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.625108 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.625121 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630362 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630417 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630472 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630490 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630509 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5zh2\" (UniqueName: \"kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630537 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630579 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.630613 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: E1013 13:02:09.702349 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c5f4aa6_d9cf_48ba_8331_9ecd76c24783.slice\": RecentStats: unable to find data in memory cache]" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732568 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732775 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732845 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732885 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732915 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5zh2\" (UniqueName: \"kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.732970 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.733133 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.734758 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.734950 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.737435 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.737846 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.738321 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.738665 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.742028 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.749696 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5zh2\" (UniqueName: \"kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2\") pod \"ceilometer-0\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " pod="openstack/ceilometer-0" Oct 13 13:02:09 crc kubenswrapper[4678]: I1013 13:02:09.945356 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.111790 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.138853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle\") pod \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.138911 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7kbn\" (UniqueName: \"kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn\") pod \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.138952 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data\") pod \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\" (UID: \"173b4d26-e8d5-4ee8-975a-3b133c992d6e\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.143922 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn" (OuterVolumeSpecName: "kube-api-access-g7kbn") pod "173b4d26-e8d5-4ee8-975a-3b133c992d6e" (UID: "173b4d26-e8d5-4ee8-975a-3b133c992d6e"). InnerVolumeSpecName "kube-api-access-g7kbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.188585 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data" (OuterVolumeSpecName: "config-data") pod "173b4d26-e8d5-4ee8-975a-3b133c992d6e" (UID: "173b4d26-e8d5-4ee8-975a-3b133c992d6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.189014 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "173b4d26-e8d5-4ee8-975a-3b133c992d6e" (UID: "173b4d26-e8d5-4ee8-975a-3b133c992d6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.190653 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.223647 4678 generic.go:334] "Generic (PLEG): container finished" podID="893c2b61-85ae-48d0-9914-b85046508949" containerID="f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d" exitCode=137 Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.223713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerDied","Data":"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d"} Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.223744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"893c2b61-85ae-48d0-9914-b85046508949","Type":"ContainerDied","Data":"90ace4eeecacaa2ecb59f601b2c91ebcc1d98fd148bd31ddb67c7e3660e302b3"} Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.223763 4678 scope.go:117] "RemoveContainer" containerID="f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.223907 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.234513 4678 generic.go:334] "Generic (PLEG): container finished" podID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" containerID="6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b" exitCode=137 Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.234664 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.235996 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"173b4d26-e8d5-4ee8-975a-3b133c992d6e","Type":"ContainerDied","Data":"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b"} Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.236045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"173b4d26-e8d5-4ee8-975a-3b133c992d6e","Type":"ContainerDied","Data":"e3ad0256838a96d2528e9858076e0cb2b79e24b0f1c6d905cd4971e568295a80"} Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.242299 4678 generic.go:334] "Generic (PLEG): container finished" podID="40584b7d-1111-4a48-afdb-0b76eb151195" containerID="2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1" exitCode=143 Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.242315 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerDied","Data":"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1"} Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.243155 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle\") pod \"893c2b61-85ae-48d0-9914-b85046508949\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.243223 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs\") pod \"893c2b61-85ae-48d0-9914-b85046508949\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.243358 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd7bk\" (UniqueName: \"kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk\") pod \"893c2b61-85ae-48d0-9914-b85046508949\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.243646 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data\") pod \"893c2b61-85ae-48d0-9914-b85046508949\" (UID: \"893c2b61-85ae-48d0-9914-b85046508949\") " Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.243725 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs" (OuterVolumeSpecName: "logs") pod "893c2b61-85ae-48d0-9914-b85046508949" (UID: "893c2b61-85ae-48d0-9914-b85046508949"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.244200 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.244219 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7kbn\" (UniqueName: \"kubernetes.io/projected/173b4d26-e8d5-4ee8-975a-3b133c992d6e-kube-api-access-g7kbn\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.244232 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/173b4d26-e8d5-4ee8-975a-3b133c992d6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.244243 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/893c2b61-85ae-48d0-9914-b85046508949-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.251104 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk" (OuterVolumeSpecName: "kube-api-access-jd7bk") pod "893c2b61-85ae-48d0-9914-b85046508949" (UID: "893c2b61-85ae-48d0-9914-b85046508949"). InnerVolumeSpecName "kube-api-access-jd7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.269069 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "893c2b61-85ae-48d0-9914-b85046508949" (UID: "893c2b61-85ae-48d0-9914-b85046508949"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.271981 4678 scope.go:117] "RemoveContainer" containerID="948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.285075 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.289442 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data" (OuterVolumeSpecName: "config-data") pod "893c2b61-85ae-48d0-9914-b85046508949" (UID: "893c2b61-85ae-48d0-9914-b85046508949"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.303531 4678 scope.go:117] "RemoveContainer" containerID="f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d" Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.304279 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d\": container with ID starting with f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d not found: ID does not exist" containerID="f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.304309 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d"} err="failed to get container status \"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d\": rpc error: code = NotFound desc = could not find container \"f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d\": container with ID starting with f68b4829c1457bf284766a05f7b752947405d8f480f7bb7cd0ae6dcddf9c9d5d not found: ID does not exist" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.304329 4678 scope.go:117] "RemoveContainer" containerID="948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55" Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.304609 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55\": container with ID starting with 948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55 not found: ID does not exist" containerID="948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.304636 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55"} err="failed to get container status \"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55\": rpc error: code = NotFound desc = could not find container \"948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55\": container with ID starting with 948643fc1135af683a37dfee2b33f418925802d7793b6e22a2a60b9181766b55 not found: ID does not exist" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.304651 4678 scope.go:117] "RemoveContainer" containerID="6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.311555 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.320911 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.321391 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-log" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321411 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-log" Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.321429 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321436 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.321446 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-metadata" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321454 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-metadata" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321627 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-metadata" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321636 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" containerName="nova-cell1-novncproxy-novncproxy" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.321651 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="893c2b61-85ae-48d0-9914-b85046508949" containerName="nova-metadata-log" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.322274 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.322814 4678 scope.go:117] "RemoveContainer" containerID="6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b" Oct 13 13:02:10 crc kubenswrapper[4678]: E1013 13:02:10.323526 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b\": container with ID starting with 6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b not found: ID does not exist" containerID="6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.323552 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b"} err="failed to get container status \"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b\": rpc error: code = NotFound desc = could not find container \"6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b\": container with ID starting with 6482cbba04d991814e080a75fb3b0d1ced62614fee34988349bfcb0e9cc6147b not found: ID does not exist" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.325235 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.325315 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.325381 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.328453 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.346085 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsght\" (UniqueName: \"kubernetes.io/projected/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-kube-api-access-bsght\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.346173 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.346284 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.346763 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.346815 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.349745 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.349779 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893c2b61-85ae-48d0-9914-b85046508949-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.349790 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd7bk\" (UniqueName: \"kubernetes.io/projected/893c2b61-85ae-48d0-9914-b85046508949-kube-api-access-jd7bk\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.374844 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.448006 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.451844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.451909 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.451938 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.452025 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsght\" (UniqueName: \"kubernetes.io/projected/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-kube-api-access-bsght\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.452042 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.456087 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.456809 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.456962 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.460198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.475175 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsght\" (UniqueName: \"kubernetes.io/projected/5b4c17c9-33f5-406f-9e2f-2fa0eef6171d-kube-api-access-bsght\") pod \"nova-cell1-novncproxy-0\" (UID: \"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.585818 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.609658 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="173b4d26-e8d5-4ee8-975a-3b133c992d6e" path="/var/lib/kubelet/pods/173b4d26-e8d5-4ee8-975a-3b133c992d6e/volumes" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.610231 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5f4aa6-d9cf-48ba-8331-9ecd76c24783" path="/var/lib/kubelet/pods/3c5f4aa6-d9cf-48ba-8331-9ecd76c24783/volumes" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.610906 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.610935 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.612437 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.615145 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.615340 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.618105 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.660748 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.762465 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.762979 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcbrm\" (UniqueName: \"kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.763162 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.763290 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.763361 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.865032 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.865800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcbrm\" (UniqueName: \"kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.865834 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.865898 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.865941 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.866514 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.871245 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.876733 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.878894 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.888717 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcbrm\" (UniqueName: \"kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm\") pod \"nova-metadata-0\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " pod="openstack/nova-metadata-0" Oct 13 13:02:10 crc kubenswrapper[4678]: I1013 13:02:10.980976 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:11 crc kubenswrapper[4678]: I1013 13:02:11.093359 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 13 13:02:11 crc kubenswrapper[4678]: I1013 13:02:11.263702 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerStarted","Data":"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3"} Oct 13 13:02:11 crc kubenswrapper[4678]: I1013 13:02:11.263740 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerStarted","Data":"d5005ea01b72e14521ec98922c4e4dc196ec8b7a80203f6e8f1d1d42069ad0b5"} Oct 13 13:02:11 crc kubenswrapper[4678]: I1013 13:02:11.269371 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d","Type":"ContainerStarted","Data":"00b33dfd9280d92380717b6e14d3868a77489b4924db83ed8600232702723df3"} Oct 13 13:02:11 crc kubenswrapper[4678]: I1013 13:02:11.392660 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:11 crc kubenswrapper[4678]: W1013 13:02:11.397437 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb79ca448_92b8_4890_9840_bcc2bb7a4cc6.slice/crio-1dd96d5bc868e4cad87e3dc281fb856461ae8fc853c61ef5f09263b5bb546f22 WatchSource:0}: Error finding container 1dd96d5bc868e4cad87e3dc281fb856461ae8fc853c61ef5f09263b5bb546f22: Status 404 returned error can't find the container with id 1dd96d5bc868e4cad87e3dc281fb856461ae8fc853c61ef5f09263b5bb546f22 Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.281380 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerStarted","Data":"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b"} Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.281939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerStarted","Data":"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b"} Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.281956 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerStarted","Data":"1dd96d5bc868e4cad87e3dc281fb856461ae8fc853c61ef5f09263b5bb546f22"} Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.283622 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5b4c17c9-33f5-406f-9e2f-2fa0eef6171d","Type":"ContainerStarted","Data":"942ea24b058a845654dc03b39f662acc199cc743dadadf8816d1fc38ec7143d7"} Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.285745 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerStarted","Data":"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba"} Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.303074 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.3016044239999998 podStartE2EDuration="2.301604424s" podCreationTimestamp="2025-10-13 13:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:12.296265295 +0000 UTC m=+1120.380803179" watchObservedRunningTime="2025-10-13 13:02:12.301604424 +0000 UTC m=+1120.386142308" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.339368 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.339347347 podStartE2EDuration="2.339347347s" podCreationTimestamp="2025-10-13 13:02:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:12.331215225 +0000 UTC m=+1120.415753109" watchObservedRunningTime="2025-10-13 13:02:12.339347347 +0000 UTC m=+1120.423885231" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.615410 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="893c2b61-85ae-48d0-9914-b85046508949" path="/var/lib/kubelet/pods/893c2b61-85ae-48d0-9914-b85046508949/volumes" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.790108 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.914099 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data\") pod \"40584b7d-1111-4a48-afdb-0b76eb151195\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.914468 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle\") pod \"40584b7d-1111-4a48-afdb-0b76eb151195\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.914867 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs\") pod \"40584b7d-1111-4a48-afdb-0b76eb151195\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.915033 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49xxl\" (UniqueName: \"kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl\") pod \"40584b7d-1111-4a48-afdb-0b76eb151195\" (UID: \"40584b7d-1111-4a48-afdb-0b76eb151195\") " Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.915518 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs" (OuterVolumeSpecName: "logs") pod "40584b7d-1111-4a48-afdb-0b76eb151195" (UID: "40584b7d-1111-4a48-afdb-0b76eb151195"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.918949 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl" (OuterVolumeSpecName: "kube-api-access-49xxl") pod "40584b7d-1111-4a48-afdb-0b76eb151195" (UID: "40584b7d-1111-4a48-afdb-0b76eb151195"). InnerVolumeSpecName "kube-api-access-49xxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.943679 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data" (OuterVolumeSpecName: "config-data") pod "40584b7d-1111-4a48-afdb-0b76eb151195" (UID: "40584b7d-1111-4a48-afdb-0b76eb151195"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:12 crc kubenswrapper[4678]: I1013 13:02:12.963595 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40584b7d-1111-4a48-afdb-0b76eb151195" (UID: "40584b7d-1111-4a48-afdb-0b76eb151195"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.017656 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.017692 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40584b7d-1111-4a48-afdb-0b76eb151195-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.017703 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49xxl\" (UniqueName: \"kubernetes.io/projected/40584b7d-1111-4a48-afdb-0b76eb151195-kube-api-access-49xxl\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.017715 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40584b7d-1111-4a48-afdb-0b76eb151195-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.296451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerStarted","Data":"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c"} Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.298482 4678 generic.go:334] "Generic (PLEG): container finished" podID="40584b7d-1111-4a48-afdb-0b76eb151195" containerID="656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3" exitCode=0 Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.298526 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.298569 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerDied","Data":"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3"} Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.298604 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"40584b7d-1111-4a48-afdb-0b76eb151195","Type":"ContainerDied","Data":"c32c4ea52963b1c17b2ddefb93d45fb60b97ff3452b983d89cc31168b6335536"} Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.298625 4678 scope.go:117] "RemoveContainer" containerID="656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.346147 4678 scope.go:117] "RemoveContainer" containerID="2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.347385 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.366282 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.372004 4678 scope.go:117] "RemoveContainer" containerID="656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3" Oct 13 13:02:13 crc kubenswrapper[4678]: E1013 13:02:13.373680 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3\": container with ID starting with 656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3 not found: ID does not exist" containerID="656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.373718 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3"} err="failed to get container status \"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3\": rpc error: code = NotFound desc = could not find container \"656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3\": container with ID starting with 656a7db897d9c95236022e0f3ae465701658b827121dad5cc5b28d4bc7ccb7f3 not found: ID does not exist" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.373742 4678 scope.go:117] "RemoveContainer" containerID="2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1" Oct 13 13:02:13 crc kubenswrapper[4678]: E1013 13:02:13.377469 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1\": container with ID starting with 2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1 not found: ID does not exist" containerID="2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.377503 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1"} err="failed to get container status \"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1\": rpc error: code = NotFound desc = could not find container \"2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1\": container with ID starting with 2fdfc917ebbc802aa8ee3626449b6a54c5071fbd522104c82651e9e01034e4f1 not found: ID does not exist" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.387978 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:13 crc kubenswrapper[4678]: E1013 13:02:13.388551 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-api" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.388579 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-api" Oct 13 13:02:13 crc kubenswrapper[4678]: E1013 13:02:13.388608 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-log" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.388616 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-log" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.388859 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-log" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.388895 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" containerName="nova-api-api" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.390175 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.393323 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.393432 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.396935 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.398973 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.446922 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.446967 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.446986 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.447017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.447233 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.447624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmnvc\" (UniqueName: \"kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549716 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmnvc\" (UniqueName: \"kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549811 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549868 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549900 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.549925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.550525 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.554504 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.555235 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.555441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.556511 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.567382 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmnvc\" (UniqueName: \"kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc\") pod \"nova-api-0\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " pod="openstack/nova-api-0" Oct 13 13:02:13 crc kubenswrapper[4678]: I1013 13:02:13.720831 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.191335 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.314744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerStarted","Data":"0f43915091653d1d03ee38e7ef66093cef2bcc3ce85a3c1c761795e0725299ce"} Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318240 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerStarted","Data":"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f"} Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318348 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318383 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-central-agent" containerID="cri-o://131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3" gracePeriod=30 Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318424 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="proxy-httpd" containerID="cri-o://9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f" gracePeriod=30 Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318401 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="sg-core" containerID="cri-o://a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c" gracePeriod=30 Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.318571 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-notification-agent" containerID="cri-o://9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba" gracePeriod=30 Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.375558 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.854002515 podStartE2EDuration="5.375529776s" podCreationTimestamp="2025-10-13 13:02:09 +0000 UTC" firstStartedPulling="2025-10-13 13:02:10.456076038 +0000 UTC m=+1118.540613952" lastFinishedPulling="2025-10-13 13:02:13.977603329 +0000 UTC m=+1122.062141213" observedRunningTime="2025-10-13 13:02:14.352400544 +0000 UTC m=+1122.436938468" watchObservedRunningTime="2025-10-13 13:02:14.375529776 +0000 UTC m=+1122.460067690" Oct 13 13:02:14 crc kubenswrapper[4678]: I1013 13:02:14.604692 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40584b7d-1111-4a48-afdb-0b76eb151195" path="/var/lib/kubelet/pods/40584b7d-1111-4a48-afdb-0b76eb151195/volumes" Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.348801 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerStarted","Data":"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650"} Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.349208 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerStarted","Data":"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676"} Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.357989 4678 generic.go:334] "Generic (PLEG): container finished" podID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerID="9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f" exitCode=0 Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.358027 4678 generic.go:334] "Generic (PLEG): container finished" podID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerID="a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c" exitCode=2 Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.358040 4678 generic.go:334] "Generic (PLEG): container finished" podID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerID="9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba" exitCode=0 Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.358083 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerDied","Data":"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f"} Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.358111 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerDied","Data":"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c"} Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.358125 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerDied","Data":"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba"} Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.383213 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.383183614 podStartE2EDuration="2.383183614s" podCreationTimestamp="2025-10-13 13:02:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:15.367503446 +0000 UTC m=+1123.452041340" watchObservedRunningTime="2025-10-13 13:02:15.383183614 +0000 UTC m=+1123.467721548" Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.662157 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.981756 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:02:15 crc kubenswrapper[4678]: I1013 13:02:15.981829 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.346905 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.376704 4678 generic.go:334] "Generic (PLEG): container finished" podID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerID="131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3" exitCode=0 Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.376795 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.376800 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerDied","Data":"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3"} Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.376982 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4846dd73-f310-459a-85fb-5ee51fec42a7","Type":"ContainerDied","Data":"d5005ea01b72e14521ec98922c4e4dc196ec8b7a80203f6e8f1d1d42069ad0b5"} Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.377021 4678 scope.go:117] "RemoveContainer" containerID="9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.404565 4678 scope.go:117] "RemoveContainer" containerID="a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.415994 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416132 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416202 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416418 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5zh2\" (UniqueName: \"kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416492 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416521 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle\") pod \"4846dd73-f310-459a-85fb-5ee51fec42a7\" (UID: \"4846dd73-f310-459a-85fb-5ee51fec42a7\") " Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416566 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.416987 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.417009 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4846dd73-f310-459a-85fb-5ee51fec42a7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.428331 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts" (OuterVolumeSpecName: "scripts") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.428616 4678 scope.go:117] "RemoveContainer" containerID="9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.432354 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2" (OuterVolumeSpecName: "kube-api-access-j5zh2") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "kube-api-access-j5zh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.444319 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.473586 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.500848 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.518068 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5zh2\" (UniqueName: \"kubernetes.io/projected/4846dd73-f310-459a-85fb-5ee51fec42a7-kube-api-access-j5zh2\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.518093 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.518104 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.518112 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.518120 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.519697 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data" (OuterVolumeSpecName: "config-data") pod "4846dd73-f310-459a-85fb-5ee51fec42a7" (UID: "4846dd73-f310-459a-85fb-5ee51fec42a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.553209 4678 scope.go:117] "RemoveContainer" containerID="131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.571539 4678 scope.go:117] "RemoveContainer" containerID="9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.572012 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f\": container with ID starting with 9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f not found: ID does not exist" containerID="9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.572101 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f"} err="failed to get container status \"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f\": rpc error: code = NotFound desc = could not find container \"9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f\": container with ID starting with 9815ad8822980ad5414620846df4b63cec78d64e8a0c33e3081109bc910f9c6f not found: ID does not exist" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.572134 4678 scope.go:117] "RemoveContainer" containerID="a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.572508 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c\": container with ID starting with a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c not found: ID does not exist" containerID="a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.572640 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c"} err="failed to get container status \"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c\": rpc error: code = NotFound desc = could not find container \"a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c\": container with ID starting with a66e8f76240a3d1caac2ab83dacef4f2ddee8e49ecbed3936f74c6357fe6a41c not found: ID does not exist" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.572737 4678 scope.go:117] "RemoveContainer" containerID="9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.573163 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba\": container with ID starting with 9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba not found: ID does not exist" containerID="9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.573280 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba"} err="failed to get container status \"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba\": rpc error: code = NotFound desc = could not find container \"9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba\": container with ID starting with 9ffc3ea98e552a0d081ca0ef07f37f50a726190a93af9f80b8a6608eb3599fba not found: ID does not exist" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.573381 4678 scope.go:117] "RemoveContainer" containerID="131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.573845 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3\": container with ID starting with 131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3 not found: ID does not exist" containerID="131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.573954 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3"} err="failed to get container status \"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3\": rpc error: code = NotFound desc = could not find container \"131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3\": container with ID starting with 131790f59415fe56b1988c860f92bafe897914006797aa9b63beb24c1daa4ae3 not found: ID does not exist" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.619887 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4846dd73-f310-459a-85fb-5ee51fec42a7-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.703742 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.720081 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.731170 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.731574 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="sg-core" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.731624 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="sg-core" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.731683 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-central-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.731692 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-central-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.731704 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-notification-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.731818 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-notification-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: E1013 13:02:16.731859 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="proxy-httpd" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.731865 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="proxy-httpd" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.732157 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-central-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.732168 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="sg-core" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.732179 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="ceilometer-notification-agent" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.732188 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" containerName="proxy-httpd" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.736146 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.738792 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.738906 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.739712 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.746409 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823017 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-config-data\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823068 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ch6v\" (UniqueName: \"kubernetes.io/projected/d7efc701-cf6f-4605-b72d-65e75bd5ce29-kube-api-access-6ch6v\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823090 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-scripts\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823117 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-run-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823215 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-log-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823232 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.823321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.848256 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.903922 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.904239 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="dnsmasq-dns" containerID="cri-o://67cee14a7b9c10957db18a2ced5a33dbc9c9f862cce7b649da2cc5958faeaa00" gracePeriod=10 Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924638 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-config-data\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ch6v\" (UniqueName: \"kubernetes.io/projected/d7efc701-cf6f-4605-b72d-65e75bd5ce29-kube-api-access-6ch6v\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924680 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-scripts\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924706 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924756 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-run-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924824 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-log-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.924841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.925457 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-run-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.925721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7efc701-cf6f-4605-b72d-65e75bd5ce29-log-httpd\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.930872 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.932081 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.940302 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-scripts\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.942141 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.950375 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ch6v\" (UniqueName: \"kubernetes.io/projected/d7efc701-cf6f-4605-b72d-65e75bd5ce29-kube-api-access-6ch6v\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:16 crc kubenswrapper[4678]: I1013 13:02:16.956877 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7efc701-cf6f-4605-b72d-65e75bd5ce29-config-data\") pod \"ceilometer-0\" (UID: \"d7efc701-cf6f-4605-b72d-65e75bd5ce29\") " pod="openstack/ceilometer-0" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.053349 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.391346 4678 generic.go:334] "Generic (PLEG): container finished" podID="73fbb203-e0e8-46aa-b47d-646507684992" containerID="67cee14a7b9c10957db18a2ced5a33dbc9c9f862cce7b649da2cc5958faeaa00" exitCode=0 Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.391628 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" event={"ID":"73fbb203-e0e8-46aa-b47d-646507684992","Type":"ContainerDied","Data":"67cee14a7b9c10957db18a2ced5a33dbc9c9f862cce7b649da2cc5958faeaa00"} Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.391798 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" event={"ID":"73fbb203-e0e8-46aa-b47d-646507684992","Type":"ContainerDied","Data":"e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826"} Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.391813 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3bd2f7b449bfee82ea0b470623f8103958eec7ab586a63a0b6fbc9b061e9826" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.397379 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.439811 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.439945 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.439999 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.440033 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hwmb\" (UniqueName: \"kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.440092 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.440118 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb\") pod \"73fbb203-e0e8-46aa-b47d-646507684992\" (UID: \"73fbb203-e0e8-46aa-b47d-646507684992\") " Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.467367 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb" (OuterVolumeSpecName: "kube-api-access-2hwmb") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "kube-api-access-2hwmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.508871 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.523630 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.527506 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.553308 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.553364 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.553385 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.553396 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hwmb\" (UniqueName: \"kubernetes.io/projected/73fbb203-e0e8-46aa-b47d-646507684992-kube-api-access-2hwmb\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.553454 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config" (OuterVolumeSpecName: "config") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.582965 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "73fbb203-e0e8-46aa-b47d-646507684992" (UID: "73fbb203-e0e8-46aa-b47d-646507684992"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:02:17 crc kubenswrapper[4678]: W1013 13:02:17.585661 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7efc701_cf6f_4605_b72d_65e75bd5ce29.slice/crio-3cb04868d8104041c0f25e6055edc2bb7757d522800b25b03f6032154d918b7c WatchSource:0}: Error finding container 3cb04868d8104041c0f25e6055edc2bb7757d522800b25b03f6032154d918b7c: Status 404 returned error can't find the container with id 3cb04868d8104041c0f25e6055edc2bb7757d522800b25b03f6032154d918b7c Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.596809 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.655678 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:17 crc kubenswrapper[4678]: I1013 13:02:17.655717 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/73fbb203-e0e8-46aa-b47d-646507684992-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.408116 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7efc701-cf6f-4605-b72d-65e75bd5ce29","Type":"ContainerStarted","Data":"3cb04868d8104041c0f25e6055edc2bb7757d522800b25b03f6032154d918b7c"} Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.408141 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-g72dq" Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.463265 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.474582 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-g72dq"] Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.605231 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4846dd73-f310-459a-85fb-5ee51fec42a7" path="/var/lib/kubelet/pods/4846dd73-f310-459a-85fb-5ee51fec42a7/volumes" Oct 13 13:02:18 crc kubenswrapper[4678]: I1013 13:02:18.606675 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73fbb203-e0e8-46aa-b47d-646507684992" path="/var/lib/kubelet/pods/73fbb203-e0e8-46aa-b47d-646507684992/volumes" Oct 13 13:02:19 crc kubenswrapper[4678]: I1013 13:02:19.416932 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7efc701-cf6f-4605-b72d-65e75bd5ce29","Type":"ContainerStarted","Data":"470b8759c3f50abd1626f0ae65b8100a4896ac89c296a1f395000352b3faec80"} Oct 13 13:02:19 crc kubenswrapper[4678]: I1013 13:02:19.417295 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7efc701-cf6f-4605-b72d-65e75bd5ce29","Type":"ContainerStarted","Data":"ac3f8b77e02297080211c3730188e32dd36d278852fa1e6ef65d7a900deb7662"} Oct 13 13:02:20 crc kubenswrapper[4678]: I1013 13:02:20.430939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7efc701-cf6f-4605-b72d-65e75bd5ce29","Type":"ContainerStarted","Data":"e3a5cf0103ff59098c5b0d83b532989ff66ffed98095d7ac46f567fb21abb510"} Oct 13 13:02:20 crc kubenswrapper[4678]: I1013 13:02:20.662071 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:20 crc kubenswrapper[4678]: I1013 13:02:20.678164 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:20 crc kubenswrapper[4678]: I1013 13:02:20.982047 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 13:02:20 crc kubenswrapper[4678]: I1013 13:02:20.982117 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.478709 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.652576 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7jjdj"] Oct 13 13:02:21 crc kubenswrapper[4678]: E1013 13:02:21.653133 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="init" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.653153 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="init" Oct 13 13:02:21 crc kubenswrapper[4678]: E1013 13:02:21.653164 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="dnsmasq-dns" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.653183 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="dnsmasq-dns" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.653449 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fbb203-e0e8-46aa-b47d-646507684992" containerName="dnsmasq-dns" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.654204 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.656642 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.656833 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.676668 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jjdj"] Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.741403 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.741466 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.741502 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.741582 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrhr9\" (UniqueName: \"kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.843814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrhr9\" (UniqueName: \"kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.843901 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.843935 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.843966 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.848153 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.848381 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.848620 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.859602 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrhr9\" (UniqueName: \"kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9\") pod \"nova-cell1-cell-mapping-7jjdj\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.980143 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.999264 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:21 crc kubenswrapper[4678]: I1013 13:02:21.999318 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:22 crc kubenswrapper[4678]: I1013 13:02:22.439734 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jjdj"] Oct 13 13:02:22 crc kubenswrapper[4678]: I1013 13:02:22.460930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jjdj" event={"ID":"2afc7094-0e68-469c-877a-5ce77ce47807","Type":"ContainerStarted","Data":"6c4dfed708b6eda823e312d0debb4e8a76ae7f8b10a6596d265e77f6cc73e6d0"} Oct 13 13:02:22 crc kubenswrapper[4678]: I1013 13:02:22.463853 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7efc701-cf6f-4605-b72d-65e75bd5ce29","Type":"ContainerStarted","Data":"dfd1d874796cd07ef37be62e2d42ac9dcb48391334f9bad451ea95b4d4b6b46f"} Oct 13 13:02:22 crc kubenswrapper[4678]: I1013 13:02:22.496329 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.347510872 podStartE2EDuration="6.496309689s" podCreationTimestamp="2025-10-13 13:02:16 +0000 UTC" firstStartedPulling="2025-10-13 13:02:17.58864002 +0000 UTC m=+1125.673177904" lastFinishedPulling="2025-10-13 13:02:21.737438837 +0000 UTC m=+1129.821976721" observedRunningTime="2025-10-13 13:02:22.487592972 +0000 UTC m=+1130.572130876" watchObservedRunningTime="2025-10-13 13:02:22.496309689 +0000 UTC m=+1130.580847563" Oct 13 13:02:23 crc kubenswrapper[4678]: I1013 13:02:23.479810 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jjdj" event={"ID":"2afc7094-0e68-469c-877a-5ce77ce47807","Type":"ContainerStarted","Data":"4f7c25f8c1f5e4ca130fbef0a63a51f0c96899dc07dd9141356fada68d354bf7"} Oct 13 13:02:23 crc kubenswrapper[4678]: I1013 13:02:23.480698 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 13 13:02:23 crc kubenswrapper[4678]: I1013 13:02:23.721560 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:02:23 crc kubenswrapper[4678]: I1013 13:02:23.721609 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:02:24 crc kubenswrapper[4678]: I1013 13:02:24.736229 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:24 crc kubenswrapper[4678]: I1013 13:02:24.736302 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:25 crc kubenswrapper[4678]: I1013 13:02:25.505911 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:02:25 crc kubenswrapper[4678]: I1013 13:02:25.506025 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:02:25 crc kubenswrapper[4678]: I1013 13:02:25.506238 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:02:25 crc kubenswrapper[4678]: I1013 13:02:25.508124 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:02:25 crc kubenswrapper[4678]: I1013 13:02:25.508548 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06" gracePeriod=600 Oct 13 13:02:26 crc kubenswrapper[4678]: I1013 13:02:26.512575 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06" exitCode=0 Oct 13 13:02:26 crc kubenswrapper[4678]: I1013 13:02:26.512668 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06"} Oct 13 13:02:26 crc kubenswrapper[4678]: I1013 13:02:26.513186 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7"} Oct 13 13:02:26 crc kubenswrapper[4678]: I1013 13:02:26.513211 4678 scope.go:117] "RemoveContainer" containerID="ba5fcd2922d535a3bbd7630dd9ce140072cb5e28adf33883b5aba4b50e553470" Oct 13 13:02:26 crc kubenswrapper[4678]: I1013 13:02:26.538351 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7jjdj" podStartSLOduration=5.538335198 podStartE2EDuration="5.538335198s" podCreationTimestamp="2025-10-13 13:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:23.498448744 +0000 UTC m=+1131.582986658" watchObservedRunningTime="2025-10-13 13:02:26.538335198 +0000 UTC m=+1134.622873072" Oct 13 13:02:27 crc kubenswrapper[4678]: I1013 13:02:27.527997 4678 generic.go:334] "Generic (PLEG): container finished" podID="2afc7094-0e68-469c-877a-5ce77ce47807" containerID="4f7c25f8c1f5e4ca130fbef0a63a51f0c96899dc07dd9141356fada68d354bf7" exitCode=0 Oct 13 13:02:27 crc kubenswrapper[4678]: I1013 13:02:27.528186 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jjdj" event={"ID":"2afc7094-0e68-469c-877a-5ce77ce47807","Type":"ContainerDied","Data":"4f7c25f8c1f5e4ca130fbef0a63a51f0c96899dc07dd9141356fada68d354bf7"} Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.013190 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.200546 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrhr9\" (UniqueName: \"kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9\") pod \"2afc7094-0e68-469c-877a-5ce77ce47807\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.201255 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle\") pod \"2afc7094-0e68-469c-877a-5ce77ce47807\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.201503 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data\") pod \"2afc7094-0e68-469c-877a-5ce77ce47807\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.201645 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts\") pod \"2afc7094-0e68-469c-877a-5ce77ce47807\" (UID: \"2afc7094-0e68-469c-877a-5ce77ce47807\") " Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.205823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts" (OuterVolumeSpecName: "scripts") pod "2afc7094-0e68-469c-877a-5ce77ce47807" (UID: "2afc7094-0e68-469c-877a-5ce77ce47807"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.208230 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9" (OuterVolumeSpecName: "kube-api-access-nrhr9") pod "2afc7094-0e68-469c-877a-5ce77ce47807" (UID: "2afc7094-0e68-469c-877a-5ce77ce47807"). InnerVolumeSpecName "kube-api-access-nrhr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.232355 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data" (OuterVolumeSpecName: "config-data") pod "2afc7094-0e68-469c-877a-5ce77ce47807" (UID: "2afc7094-0e68-469c-877a-5ce77ce47807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.254858 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2afc7094-0e68-469c-877a-5ce77ce47807" (UID: "2afc7094-0e68-469c-877a-5ce77ce47807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.303532 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.303563 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-scripts\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.303572 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrhr9\" (UniqueName: \"kubernetes.io/projected/2afc7094-0e68-469c-877a-5ce77ce47807-kube-api-access-nrhr9\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.303582 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2afc7094-0e68-469c-877a-5ce77ce47807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.556411 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jjdj" event={"ID":"2afc7094-0e68-469c-877a-5ce77ce47807","Type":"ContainerDied","Data":"6c4dfed708b6eda823e312d0debb4e8a76ae7f8b10a6596d265e77f6cc73e6d0"} Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.556474 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c4dfed708b6eda823e312d0debb4e8a76ae7f8b10a6596d265e77f6cc73e6d0" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.556549 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jjdj" Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.726728 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.726977 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1be79df5-72d5-4616-88e5-62eda59fc6e8" containerName="nova-scheduler-scheduler" containerID="cri-o://6e245d433f359286cf38e289d8601e898386d38680a2e55ba65507c557ca5cc3" gracePeriod=30 Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.740432 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.740722 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-log" containerID="cri-o://4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676" gracePeriod=30 Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.740845 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-api" containerID="cri-o://8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650" gracePeriod=30 Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.757805 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.758030 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-log" containerID="cri-o://0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b" gracePeriod=30 Oct 13 13:02:29 crc kubenswrapper[4678]: I1013 13:02:29.758178 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-metadata" containerID="cri-o://e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b" gracePeriod=30 Oct 13 13:02:30 crc kubenswrapper[4678]: I1013 13:02:30.566848 4678 generic.go:334] "Generic (PLEG): container finished" podID="dc061189-e404-4187-8d22-d492bcb6befc" containerID="4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676" exitCode=143 Oct 13 13:02:30 crc kubenswrapper[4678]: I1013 13:02:30.566878 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerDied","Data":"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676"} Oct 13 13:02:30 crc kubenswrapper[4678]: I1013 13:02:30.569469 4678 generic.go:334] "Generic (PLEG): container finished" podID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerID="0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b" exitCode=143 Oct 13 13:02:30 crc kubenswrapper[4678]: I1013 13:02:30.569503 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerDied","Data":"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b"} Oct 13 13:02:31 crc kubenswrapper[4678]: I1013 13:02:31.583285 4678 generic.go:334] "Generic (PLEG): container finished" podID="1be79df5-72d5-4616-88e5-62eda59fc6e8" containerID="6e245d433f359286cf38e289d8601e898386d38680a2e55ba65507c557ca5cc3" exitCode=0 Oct 13 13:02:31 crc kubenswrapper[4678]: I1013 13:02:31.583404 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1be79df5-72d5-4616-88e5-62eda59fc6e8","Type":"ContainerDied","Data":"6e245d433f359286cf38e289d8601e898386d38680a2e55ba65507c557ca5cc3"} Oct 13 13:02:31 crc kubenswrapper[4678]: I1013 13:02:31.883606 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.054277 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle\") pod \"1be79df5-72d5-4616-88e5-62eda59fc6e8\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.054325 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data\") pod \"1be79df5-72d5-4616-88e5-62eda59fc6e8\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.054349 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h48gk\" (UniqueName: \"kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk\") pod \"1be79df5-72d5-4616-88e5-62eda59fc6e8\" (UID: \"1be79df5-72d5-4616-88e5-62eda59fc6e8\") " Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.061122 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk" (OuterVolumeSpecName: "kube-api-access-h48gk") pod "1be79df5-72d5-4616-88e5-62eda59fc6e8" (UID: "1be79df5-72d5-4616-88e5-62eda59fc6e8"). InnerVolumeSpecName "kube-api-access-h48gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.086571 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1be79df5-72d5-4616-88e5-62eda59fc6e8" (UID: "1be79df5-72d5-4616-88e5-62eda59fc6e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.103391 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data" (OuterVolumeSpecName: "config-data") pod "1be79df5-72d5-4616-88e5-62eda59fc6e8" (UID: "1be79df5-72d5-4616-88e5-62eda59fc6e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.156628 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.156675 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1be79df5-72d5-4616-88e5-62eda59fc6e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.156689 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h48gk\" (UniqueName: \"kubernetes.io/projected/1be79df5-72d5-4616-88e5-62eda59fc6e8-kube-api-access-h48gk\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.599842 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.613846 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1be79df5-72d5-4616-88e5-62eda59fc6e8","Type":"ContainerDied","Data":"0ee5f704bce9c3ec9d466e46430819d3453852a164c211424604c73c54559aa4"} Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.613916 4678 scope.go:117] "RemoveContainer" containerID="6e245d433f359286cf38e289d8601e898386d38680a2e55ba65507c557ca5cc3" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.659345 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.680313 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.695579 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:32 crc kubenswrapper[4678]: E1013 13:02:32.697914 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2afc7094-0e68-469c-877a-5ce77ce47807" containerName="nova-manage" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.697951 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2afc7094-0e68-469c-877a-5ce77ce47807" containerName="nova-manage" Oct 13 13:02:32 crc kubenswrapper[4678]: E1013 13:02:32.697993 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1be79df5-72d5-4616-88e5-62eda59fc6e8" containerName="nova-scheduler-scheduler" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.698007 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1be79df5-72d5-4616-88e5-62eda59fc6e8" containerName="nova-scheduler-scheduler" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.698446 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2afc7094-0e68-469c-877a-5ce77ce47807" containerName="nova-manage" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.698480 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1be79df5-72d5-4616-88e5-62eda59fc6e8" containerName="nova-scheduler-scheduler" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.699744 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.703493 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.704700 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.871455 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dxw9\" (UniqueName: \"kubernetes.io/projected/ac735c75-e8f1-402a-a7d7-d213a9badd88-kube-api-access-7dxw9\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.871753 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.871922 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-config-data\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.973253 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dxw9\" (UniqueName: \"kubernetes.io/projected/ac735c75-e8f1-402a-a7d7-d213a9badd88-kube-api-access-7dxw9\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.973427 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.973488 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-config-data\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.978613 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-config-data\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.980895 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac735c75-e8f1-402a-a7d7-d213a9badd88-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:32 crc kubenswrapper[4678]: I1013 13:02:32.994942 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dxw9\" (UniqueName: \"kubernetes.io/projected/ac735c75-e8f1-402a-a7d7-d213a9badd88-kube-api-access-7dxw9\") pod \"nova-scheduler-0\" (UID: \"ac735c75-e8f1-402a-a7d7-d213a9badd88\") " pod="openstack/nova-scheduler-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.031231 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.217992 4678 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.368045 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.413035 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489001 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489609 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489640 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmnvc\" (UniqueName: \"kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489819 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.489876 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs\") pod \"dc061189-e404-4187-8d22-d492bcb6befc\" (UID: \"dc061189-e404-4187-8d22-d492bcb6befc\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.490702 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs" (OuterVolumeSpecName: "logs") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.493953 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc" (OuterVolumeSpecName: "kube-api-access-pmnvc") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "kube-api-access-pmnvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.523743 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.528267 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data" (OuterVolumeSpecName: "config-data") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: W1013 13:02:33.547846 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac735c75_e8f1_402a_a7d7_d213a9badd88.slice/crio-869bf41953c83236ec701377ad3ecef2d0674b25febc7f9faaf8ef0133b8c517 WatchSource:0}: Error finding container 869bf41953c83236ec701377ad3ecef2d0674b25febc7f9faaf8ef0133b8c517: Status 404 returned error can't find the container with id 869bf41953c83236ec701377ad3ecef2d0674b25febc7f9faaf8ef0133b8c517 Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.548397 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.556760 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.571414 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "dc061189-e404-4187-8d22-d492bcb6befc" (UID: "dc061189-e404-4187-8d22-d492bcb6befc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.591708 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data\") pod \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.591818 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcbrm\" (UniqueName: \"kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm\") pod \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.591879 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs\") pod \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592234 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle\") pod \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592286 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs\") pod \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\" (UID: \"b79ca448-92b8-4890-9840-bcc2bb7a4cc6\") " Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592748 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592768 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592780 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc061189-e404-4187-8d22-d492bcb6befc-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592792 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592803 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc061189-e404-4187-8d22-d492bcb6befc-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.592818 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmnvc\" (UniqueName: \"kubernetes.io/projected/dc061189-e404-4187-8d22-d492bcb6befc-kube-api-access-pmnvc\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.593353 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs" (OuterVolumeSpecName: "logs") pod "b79ca448-92b8-4890-9840-bcc2bb7a4cc6" (UID: "b79ca448-92b8-4890-9840-bcc2bb7a4cc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.594536 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm" (OuterVolumeSpecName: "kube-api-access-rcbrm") pod "b79ca448-92b8-4890-9840-bcc2bb7a4cc6" (UID: "b79ca448-92b8-4890-9840-bcc2bb7a4cc6"). InnerVolumeSpecName "kube-api-access-rcbrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.624624 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac735c75-e8f1-402a-a7d7-d213a9badd88","Type":"ContainerStarted","Data":"869bf41953c83236ec701377ad3ecef2d0674b25febc7f9faaf8ef0133b8c517"} Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.626866 4678 generic.go:334] "Generic (PLEG): container finished" podID="dc061189-e404-4187-8d22-d492bcb6befc" containerID="8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650" exitCode=0 Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.626914 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.626908 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerDied","Data":"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650"} Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.626989 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"dc061189-e404-4187-8d22-d492bcb6befc","Type":"ContainerDied","Data":"0f43915091653d1d03ee38e7ef66093cef2bcc3ce85a3c1c761795e0725299ce"} Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.627012 4678 scope.go:117] "RemoveContainer" containerID="8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.634895 4678 generic.go:334] "Generic (PLEG): container finished" podID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerID="e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b" exitCode=0 Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.634952 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerDied","Data":"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b"} Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.634979 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b79ca448-92b8-4890-9840-bcc2bb7a4cc6","Type":"ContainerDied","Data":"1dd96d5bc868e4cad87e3dc281fb856461ae8fc853c61ef5f09263b5bb546f22"} Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.635067 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.642443 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data" (OuterVolumeSpecName: "config-data") pod "b79ca448-92b8-4890-9840-bcc2bb7a4cc6" (UID: "b79ca448-92b8-4890-9840-bcc2bb7a4cc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.653516 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b79ca448-92b8-4890-9840-bcc2bb7a4cc6" (UID: "b79ca448-92b8-4890-9840-bcc2bb7a4cc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.657110 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b79ca448-92b8-4890-9840-bcc2bb7a4cc6" (UID: "b79ca448-92b8-4890-9840-bcc2bb7a4cc6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.694914 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-logs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.694944 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.695288 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcbrm\" (UniqueName: \"kubernetes.io/projected/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-kube-api-access-rcbrm\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.695305 4678 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.695318 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79ca448-92b8-4890-9840-bcc2bb7a4cc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.712284 4678 scope.go:117] "RemoveContainer" containerID="4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.746599 4678 scope.go:117] "RemoveContainer" containerID="8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.747511 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650\": container with ID starting with 8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650 not found: ID does not exist" containerID="8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.747547 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650"} err="failed to get container status \"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650\": rpc error: code = NotFound desc = could not find container \"8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650\": container with ID starting with 8889aba9714bca4cdf6c2edba65662bcaa8c072f2b12d796ef879f12207ef650 not found: ID does not exist" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.747571 4678 scope.go:117] "RemoveContainer" containerID="4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.747894 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676\": container with ID starting with 4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676 not found: ID does not exist" containerID="4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.747924 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676"} err="failed to get container status \"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676\": rpc error: code = NotFound desc = could not find container \"4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676\": container with ID starting with 4f99468825f7cef8b16abfdce49cc6208278f17bd192da44354a59d7ca42a676 not found: ID does not exist" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.747941 4678 scope.go:117] "RemoveContainer" containerID="e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.752248 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.767354 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.778904 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.779418 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-log" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779439 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-log" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.779460 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-metadata" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779471 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-metadata" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.779505 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-api" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779516 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-api" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.779539 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-log" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779551 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-log" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779849 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-log" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779872 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-log" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779893 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" containerName="nova-metadata-metadata" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.779908 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc061189-e404-4187-8d22-d492bcb6befc" containerName="nova-api-api" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.781088 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.784272 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.787554 4678 scope.go:117] "RemoveContainer" containerID="0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.787813 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.788381 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.789202 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.835292 4678 scope.go:117] "RemoveContainer" containerID="e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837340 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-config-data\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837421 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh4x\" (UniqueName: \"kubernetes.io/projected/6c38d115-b798-4e2c-986c-c9541832f1bb-kube-api-access-4zh4x\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837503 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-public-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837610 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.837701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c38d115-b798-4e2c-986c-c9541832f1bb-logs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.843299 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b\": container with ID starting with e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b not found: ID does not exist" containerID="e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.843366 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b"} err="failed to get container status \"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b\": rpc error: code = NotFound desc = could not find container \"e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b\": container with ID starting with e179d2ac07a33c64fc2c6e9359b700374fd8ae09a4a6a10698380d483ddd2e0b not found: ID does not exist" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.843395 4678 scope.go:117] "RemoveContainer" containerID="0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b" Oct 13 13:02:33 crc kubenswrapper[4678]: E1013 13:02:33.846184 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b\": container with ID starting with 0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b not found: ID does not exist" containerID="0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.846318 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b"} err="failed to get container status \"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b\": rpc error: code = NotFound desc = could not find container \"0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b\": container with ID starting with 0a1474af968a2df558885f568dd4895a16f32472be915d2cf73d2a6c952d455b not found: ID does not exist" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.940212 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.940581 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.940713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c38d115-b798-4e2c-986c-c9541832f1bb-logs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.940872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-config-data\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.941010 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh4x\" (UniqueName: \"kubernetes.io/projected/6c38d115-b798-4e2c-986c-c9541832f1bb-kube-api-access-4zh4x\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.941199 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-public-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.944866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.945198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-public-tls-certs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.945263 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.946463 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c38d115-b798-4e2c-986c-c9541832f1bb-config-data\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.951146 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c38d115-b798-4e2c-986c-c9541832f1bb-logs\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.973109 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.975872 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh4x\" (UniqueName: \"kubernetes.io/projected/6c38d115-b798-4e2c-986c-c9541832f1bb-kube-api-access-4zh4x\") pod \"nova-api-0\" (UID: \"6c38d115-b798-4e2c-986c-c9541832f1bb\") " pod="openstack/nova-api-0" Oct 13 13:02:33 crc kubenswrapper[4678]: I1013 13:02:33.982620 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.004322 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.006426 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.009879 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.010170 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.023914 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.043547 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.043701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b7bb30d-3831-467e-8067-67626163b30a-logs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.043827 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.043874 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxzn\" (UniqueName: \"kubernetes.io/projected/0b7bb30d-3831-467e-8067-67626163b30a-kube-api-access-gzxzn\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.043919 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-config-data\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.144580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b7bb30d-3831-467e-8067-67626163b30a-logs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.144678 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.144713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxzn\" (UniqueName: \"kubernetes.io/projected/0b7bb30d-3831-467e-8067-67626163b30a-kube-api-access-gzxzn\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.144742 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-config-data\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.144816 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.145855 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b7bb30d-3831-467e-8067-67626163b30a-logs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.145910 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.149986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.150484 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-config-data\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.151018 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b7bb30d-3831-467e-8067-67626163b30a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.167273 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxzn\" (UniqueName: \"kubernetes.io/projected/0b7bb30d-3831-467e-8067-67626163b30a-kube-api-access-gzxzn\") pod \"nova-metadata-0\" (UID: \"0b7bb30d-3831-467e-8067-67626163b30a\") " pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.340844 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.610187 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1be79df5-72d5-4616-88e5-62eda59fc6e8" path="/var/lib/kubelet/pods/1be79df5-72d5-4616-88e5-62eda59fc6e8/volumes" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.610877 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79ca448-92b8-4890-9840-bcc2bb7a4cc6" path="/var/lib/kubelet/pods/b79ca448-92b8-4890-9840-bcc2bb7a4cc6/volumes" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.611463 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc061189-e404-4187-8d22-d492bcb6befc" path="/var/lib/kubelet/pods/dc061189-e404-4187-8d22-d492bcb6befc/volumes" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.619631 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 13 13:02:34 crc kubenswrapper[4678]: W1013 13:02:34.621804 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c38d115_b798_4e2c_986c_c9541832f1bb.slice/crio-fc82cce0cf299fc1ed655e59f6beb7faa63fd9c79bf06e13cf7bfaa0dc8d8006 WatchSource:0}: Error finding container fc82cce0cf299fc1ed655e59f6beb7faa63fd9c79bf06e13cf7bfaa0dc8d8006: Status 404 returned error can't find the container with id fc82cce0cf299fc1ed655e59f6beb7faa63fd9c79bf06e13cf7bfaa0dc8d8006 Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.644544 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ac735c75-e8f1-402a-a7d7-d213a9badd88","Type":"ContainerStarted","Data":"c838d5c83156aaff1af36949ac4ccf0aad6bf054cb9c8d548ff8bc1032b9f736"} Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.648947 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c38d115-b798-4e2c-986c-c9541832f1bb","Type":"ContainerStarted","Data":"fc82cce0cf299fc1ed655e59f6beb7faa63fd9c79bf06e13cf7bfaa0dc8d8006"} Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.660813 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.660799198 podStartE2EDuration="2.660799198s" podCreationTimestamp="2025-10-13 13:02:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:34.656623119 +0000 UTC m=+1142.741161023" watchObservedRunningTime="2025-10-13 13:02:34.660799198 +0000 UTC m=+1142.745337082" Oct 13 13:02:34 crc kubenswrapper[4678]: I1013 13:02:34.806972 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 13 13:02:34 crc kubenswrapper[4678]: W1013 13:02:34.809225 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b7bb30d_3831_467e_8067_67626163b30a.slice/crio-2a891e305fa1401c42c94900e69d7412a3e22d390c7974a12852ccfffc246f33 WatchSource:0}: Error finding container 2a891e305fa1401c42c94900e69d7412a3e22d390c7974a12852ccfffc246f33: Status 404 returned error can't find the container with id 2a891e305fa1401c42c94900e69d7412a3e22d390c7974a12852ccfffc246f33 Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.662541 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0b7bb30d-3831-467e-8067-67626163b30a","Type":"ContainerStarted","Data":"a73818de6fa0024923dfbc0fc1a68e9a91a8fb0f3d51c80c9ceea6991c481243"} Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.662868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0b7bb30d-3831-467e-8067-67626163b30a","Type":"ContainerStarted","Data":"911ed950b9a0e182d11e5a157fdfea4ad9aab5db699b26839afa614161426563"} Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.662883 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0b7bb30d-3831-467e-8067-67626163b30a","Type":"ContainerStarted","Data":"2a891e305fa1401c42c94900e69d7412a3e22d390c7974a12852ccfffc246f33"} Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.667718 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c38d115-b798-4e2c-986c-c9541832f1bb","Type":"ContainerStarted","Data":"ee3f88d4f07ece9c252883f755549eb87a8643e78f10f3e95de41b196f72cd45"} Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.667774 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6c38d115-b798-4e2c-986c-c9541832f1bb","Type":"ContainerStarted","Data":"3ff9644992d5eb22c1ad2237c7bbdc64ccb312806f6a9d5a4b111e1ef508c4de"} Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.689988 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.689973995 podStartE2EDuration="2.689973995s" podCreationTimestamp="2025-10-13 13:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:35.689217496 +0000 UTC m=+1143.773755390" watchObservedRunningTime="2025-10-13 13:02:35.689973995 +0000 UTC m=+1143.774511879" Oct 13 13:02:35 crc kubenswrapper[4678]: I1013 13:02:35.734986 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.734967056 podStartE2EDuration="2.734967056s" podCreationTimestamp="2025-10-13 13:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:02:35.725358696 +0000 UTC m=+1143.809896610" watchObservedRunningTime="2025-10-13 13:02:35.734967056 +0000 UTC m=+1143.819504950" Oct 13 13:02:38 crc kubenswrapper[4678]: I1013 13:02:38.031423 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 13 13:02:39 crc kubenswrapper[4678]: I1013 13:02:39.341718 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:02:39 crc kubenswrapper[4678]: I1013 13:02:39.342172 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 13 13:02:43 crc kubenswrapper[4678]: I1013 13:02:43.032592 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 13 13:02:43 crc kubenswrapper[4678]: I1013 13:02:43.077584 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 13 13:02:43 crc kubenswrapper[4678]: I1013 13:02:43.801884 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 13 13:02:44 crc kubenswrapper[4678]: I1013 13:02:44.146917 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:02:44 crc kubenswrapper[4678]: I1013 13:02:44.147231 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 13 13:02:44 crc kubenswrapper[4678]: I1013 13:02:44.341435 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 13:02:44 crc kubenswrapper[4678]: I1013 13:02:44.341506 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 13 13:02:45 crc kubenswrapper[4678]: I1013 13:02:45.164203 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c38d115-b798-4e2c-986c-c9541832f1bb" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:45 crc kubenswrapper[4678]: I1013 13:02:45.164226 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6c38d115-b798-4e2c-986c-c9541832f1bb" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:45 crc kubenswrapper[4678]: I1013 13:02:45.354198 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0b7bb30d-3831-467e-8067-67626163b30a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:45 crc kubenswrapper[4678]: I1013 13:02:45.354294 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0b7bb30d-3831-467e-8067-67626163b30a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 13 13:02:47 crc kubenswrapper[4678]: I1013 13:02:47.066631 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.161159 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.162141 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.162845 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.162914 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.174140 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.175796 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.358250 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.368177 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.368970 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 13 13:02:54 crc kubenswrapper[4678]: I1013 13:02:54.906456 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 13 13:03:03 crc kubenswrapper[4678]: I1013 13:03:03.242630 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:03 crc kubenswrapper[4678]: I1013 13:03:03.977481 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:07 crc kubenswrapper[4678]: I1013 13:03:07.237181 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="rabbitmq" containerID="cri-o://dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5" gracePeriod=604797 Oct 13 13:03:07 crc kubenswrapper[4678]: I1013 13:03:07.765960 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 13 13:03:08 crc kubenswrapper[4678]: I1013 13:03:08.348482 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="rabbitmq" containerID="cri-o://7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91" gracePeriod=604796 Oct 13 13:03:13 crc kubenswrapper[4678]: I1013 13:03:13.927162 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.000855 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.000956 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.000982 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001007 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001036 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001072 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001099 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45gf8\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001125 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001171 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001196 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001530 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001560 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.001802 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf\") pod \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\" (UID: \"b0795d10-95f2-4bf0-b15c-863ed19b32cd\") " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.002504 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.002521 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.009757 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.010106 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8" (OuterVolumeSpecName: "kube-api-access-45gf8") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "kube-api-access-45gf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.016385 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info" (OuterVolumeSpecName: "pod-info") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.018791 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.022262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.031774 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.051296 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data" (OuterVolumeSpecName: "config-data") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.093955 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf" (OuterVolumeSpecName: "server-conf") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.095993 4678 generic.go:334] "Generic (PLEG): container finished" podID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerID="dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5" exitCode=0 Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.096043 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerDied","Data":"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5"} Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.096086 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b0795d10-95f2-4bf0-b15c-863ed19b32cd","Type":"ContainerDied","Data":"84fc16241189c25a6307c4d3e3089de91abcf1b3066c93eeaaa9c2a84769ac11"} Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.096103 4678 scope.go:117] "RemoveContainer" containerID="dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.096113 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105329 4678 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-server-conf\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105359 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105371 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105382 4678 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b0795d10-95f2-4bf0-b15c-863ed19b32cd-pod-info\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105411 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105420 4678 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b0795d10-95f2-4bf0-b15c-863ed19b32cd-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105432 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45gf8\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-kube-api-access-45gf8\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.105441 4678 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b0795d10-95f2-4bf0-b15c-863ed19b32cd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.132817 4678 scope.go:117] "RemoveContainer" containerID="0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.140683 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.162327 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b0795d10-95f2-4bf0-b15c-863ed19b32cd" (UID: "b0795d10-95f2-4bf0-b15c-863ed19b32cd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.178238 4678 scope.go:117] "RemoveContainer" containerID="dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5" Oct 13 13:03:14 crc kubenswrapper[4678]: E1013 13:03:14.182192 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5\": container with ID starting with dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5 not found: ID does not exist" containerID="dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.182233 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5"} err="failed to get container status \"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5\": rpc error: code = NotFound desc = could not find container \"dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5\": container with ID starting with dab59cfa7b17298e88bb3b7f57bbeb930caf5893a8deea07638b38a577629ea5 not found: ID does not exist" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.182257 4678 scope.go:117] "RemoveContainer" containerID="0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a" Oct 13 13:03:14 crc kubenswrapper[4678]: E1013 13:03:14.186188 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a\": container with ID starting with 0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a not found: ID does not exist" containerID="0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.186220 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a"} err="failed to get container status \"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a\": rpc error: code = NotFound desc = could not find container \"0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a\": container with ID starting with 0950b3ea570ce545840e1ff427231899478584a07b261dd3b9c76cfac5d05d0a not found: ID does not exist" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.207193 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.207229 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b0795d10-95f2-4bf0-b15c-863ed19b32cd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.425067 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.435511 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.456022 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:14 crc kubenswrapper[4678]: E1013 13:03:14.456430 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="rabbitmq" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.456444 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="rabbitmq" Oct 13 13:03:14 crc kubenswrapper[4678]: E1013 13:03:14.456465 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="setup-container" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.456470 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="setup-container" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.456651 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" containerName="rabbitmq" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.459751 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.473766 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.474439 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.474556 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.474727 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.474947 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.475100 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlgsl" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.475204 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.477453 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511211 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511284 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511304 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511361 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511381 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511415 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511446 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74rlk\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-kube-api-access-74rlk\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511473 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511489 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511512 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.511529 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.601512 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0795d10-95f2-4bf0-b15c-863ed19b32cd" path="/var/lib/kubelet/pods/b0795d10-95f2-4bf0-b15c-863ed19b32cd/volumes" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613138 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613205 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613228 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613278 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613300 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613808 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613845 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74rlk\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-kube-api-access-74rlk\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613885 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613899 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613922 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.613935 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.614237 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-config-data\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.614437 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.614519 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.614755 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.614947 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.617929 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.618132 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.618158 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.619557 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.620441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.631523 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74rlk\" (UniqueName: \"kubernetes.io/projected/1c44f667-45a9-4e57-b8f9-ffd8da960a6e-kube-api-access-74rlk\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.655674 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"1c44f667-45a9-4e57-b8f9-ffd8da960a6e\") " pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.713189 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 13 13:03:14 crc kubenswrapper[4678]: I1013 13:03:14.857891 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921317 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921373 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921461 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921531 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921556 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921587 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921631 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921675 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x5nw\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921704 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921748 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.921772 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls\") pod \"d90b147e-003a-4981-ab10-3e933cf4be70\" (UID: \"d90b147e-003a-4981-ab10-3e933cf4be70\") " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.925234 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.927286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.929890 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.930008 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw" (OuterVolumeSpecName: "kube-api-access-5x5nw") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "kube-api-access-5x5nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.931591 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.933424 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info" (OuterVolumeSpecName: "pod-info") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.936432 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.942109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.955678 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data" (OuterVolumeSpecName: "config-data") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:14.980353 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf" (OuterVolumeSpecName: "server-conf") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023338 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023381 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023391 4678 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d90b147e-003a-4981-ab10-3e933cf4be70-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023400 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023410 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x5nw\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-kube-api-access-5x5nw\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023418 4678 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d90b147e-003a-4981-ab10-3e933cf4be70-pod-info\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023426 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023433 4678 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-server-conf\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023441 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.023449 4678 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d90b147e-003a-4981-ab10-3e933cf4be70-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.046860 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.059682 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d90b147e-003a-4981-ab10-3e933cf4be70" (UID: "d90b147e-003a-4981-ab10-3e933cf4be70"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.106673 4678 generic.go:334] "Generic (PLEG): container finished" podID="d90b147e-003a-4981-ab10-3e933cf4be70" containerID="7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91" exitCode=0 Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.106730 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerDied","Data":"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91"} Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.106725 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.106753 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d90b147e-003a-4981-ab10-3e933cf4be70","Type":"ContainerDied","Data":"23ed0708098bd2dc31620ca16f3efa29d528b78602a995a6e3a94f93a2cd5692"} Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.106771 4678 scope.go:117] "RemoveContainer" containerID="7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.124815 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d90b147e-003a-4981-ab10-3e933cf4be70-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.124836 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.132598 4678 scope.go:117] "RemoveContainer" containerID="a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.142940 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.153019 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.163305 4678 scope.go:117] "RemoveContainer" containerID="7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91" Oct 13 13:03:15 crc kubenswrapper[4678]: E1013 13:03:15.163805 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91\": container with ID starting with 7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91 not found: ID does not exist" containerID="7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.163846 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91"} err="failed to get container status \"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91\": rpc error: code = NotFound desc = could not find container \"7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91\": container with ID starting with 7b89e9ba4cbe40d743dfaabcc122739d649f653271ba7fef6b3b92e0d5c9cb91 not found: ID does not exist" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.163872 4678 scope.go:117] "RemoveContainer" containerID="a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf" Oct 13 13:03:15 crc kubenswrapper[4678]: E1013 13:03:15.164344 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf\": container with ID starting with a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf not found: ID does not exist" containerID="a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.164385 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf"} err="failed to get container status \"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf\": rpc error: code = NotFound desc = could not find container \"a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf\": container with ID starting with a0296e2f65101d83595a69848283d204664c7bc0f734f35fe70bda2f7e14c8cf not found: ID does not exist" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.172843 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: E1013 13:03:15.173415 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="rabbitmq" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.173431 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="rabbitmq" Oct 13 13:03:15 crc kubenswrapper[4678]: E1013 13:03:15.173481 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="setup-container" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.173489 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="setup-container" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.173714 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" containerName="rabbitmq" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.174963 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.178345 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.178510 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.178635 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.180362 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.180545 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zxwsg" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.180686 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.181026 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.184609 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.226782 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.226855 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.226910 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.226937 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.226999 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227028 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227065 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227087 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227174 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227236 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.227359 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh5d5\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-kube-api-access-nh5d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329621 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329691 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh5d5\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-kube-api-access-nh5d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329857 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329887 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.329908 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.330491 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.331324 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.331574 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.331907 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.332812 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.333156 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.333721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.336355 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.336677 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.340259 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.350440 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh5d5\" (UniqueName: \"kubernetes.io/projected/a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe-kube-api-access-nh5d5\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.363883 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe\") " pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.499698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.576727 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: I1013 13:03:15.974581 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 13 13:03:15 crc kubenswrapper[4678]: W1013 13:03:15.978596 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1fd79d4_f2b6_4427_9061_f56b0f1bc0fe.slice/crio-d2f2c8a89d2ddd8e7b34557b9fa68031c1877a81847262a1e0b4666090181945 WatchSource:0}: Error finding container d2f2c8a89d2ddd8e7b34557b9fa68031c1877a81847262a1e0b4666090181945: Status 404 returned error can't find the container with id d2f2c8a89d2ddd8e7b34557b9fa68031c1877a81847262a1e0b4666090181945 Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.118190 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1c44f667-45a9-4e57-b8f9-ffd8da960a6e","Type":"ContainerStarted","Data":"c3849ea8a4c0cb4105b4e966258d8972a23330fc5fc78ed3d8bc0efaaa856fea"} Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.119893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe","Type":"ContainerStarted","Data":"d2f2c8a89d2ddd8e7b34557b9fa68031c1877a81847262a1e0b4666090181945"} Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.531120 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.538438 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.542492 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.555076 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566528 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566585 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566631 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566652 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566681 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566702 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8pz2\" (UniqueName: \"kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.566725 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.609026 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d90b147e-003a-4981-ab10-3e933cf4be70" path="/var/lib/kubelet/pods/d90b147e-003a-4981-ab10-3e933cf4be70/volumes" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668513 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668564 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668629 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668654 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8pz2\" (UniqueName: \"kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668711 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668854 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.668908 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.672431 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.673326 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.675402 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.675914 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.676293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.676517 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.695282 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8pz2\" (UniqueName: \"kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2\") pod \"dnsmasq-dns-67b789f86c-bzkqx\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:16 crc kubenswrapper[4678]: I1013 13:03:16.867069 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:17 crc kubenswrapper[4678]: I1013 13:03:17.130873 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1c44f667-45a9-4e57-b8f9-ffd8da960a6e","Type":"ContainerStarted","Data":"0fb71cf7c9a1965b38ab34659773667d1c072c73b601b18a3c98dbae3096de20"} Oct 13 13:03:17 crc kubenswrapper[4678]: I1013 13:03:17.434360 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:18 crc kubenswrapper[4678]: I1013 13:03:18.144142 4678 generic.go:334] "Generic (PLEG): container finished" podID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerID="e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479" exitCode=0 Oct 13 13:03:18 crc kubenswrapper[4678]: I1013 13:03:18.144205 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" event={"ID":"13e7b0b7-853b-4d57-b959-83c1bdc457a1","Type":"ContainerDied","Data":"e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479"} Oct 13 13:03:18 crc kubenswrapper[4678]: I1013 13:03:18.144608 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" event={"ID":"13e7b0b7-853b-4d57-b959-83c1bdc457a1","Type":"ContainerStarted","Data":"27b51971375fa752335bde328aa933873e53a00bbcb14a972b46c27b4fdb91fc"} Oct 13 13:03:18 crc kubenswrapper[4678]: I1013 13:03:18.146663 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe","Type":"ContainerStarted","Data":"e8bacb69b0f4c84029a274231e9d462b7cd53a43bb56b52efbf18b83c587f533"} Oct 13 13:03:19 crc kubenswrapper[4678]: I1013 13:03:19.159506 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" event={"ID":"13e7b0b7-853b-4d57-b959-83c1bdc457a1","Type":"ContainerStarted","Data":"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe"} Oct 13 13:03:19 crc kubenswrapper[4678]: I1013 13:03:19.160286 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:19 crc kubenswrapper[4678]: I1013 13:03:19.195660 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" podStartSLOduration=3.195626444 podStartE2EDuration="3.195626444s" podCreationTimestamp="2025-10-13 13:03:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:03:19.189154288 +0000 UTC m=+1187.273692222" watchObservedRunningTime="2025-10-13 13:03:19.195626444 +0000 UTC m=+1187.280164358" Oct 13 13:03:26 crc kubenswrapper[4678]: I1013 13:03:26.869038 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:26 crc kubenswrapper[4678]: I1013 13:03:26.931739 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:03:26 crc kubenswrapper[4678]: I1013 13:03:26.931965 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="dnsmasq-dns" containerID="cri-o://4daecfe5e5fedc235149421381db7b1c6490e1e7548dbe656f57f2e1338c946e" gracePeriod=10 Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.104868 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-clzsn"] Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.106462 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.116466 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-clzsn"] Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.217348 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.217412 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.217666 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-config\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.217720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.218688 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkjw2\" (UniqueName: \"kubernetes.io/projected/d9f8446d-e7a4-46b7-b62e-f4761a396473-kube-api-access-gkjw2\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.218737 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.218850 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.237403 4678 generic.go:334] "Generic (PLEG): container finished" podID="ef9ff948-801a-4028-a003-47fbe5190158" containerID="4daecfe5e5fedc235149421381db7b1c6490e1e7548dbe656f57f2e1338c946e" exitCode=0 Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.237454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" event={"ID":"ef9ff948-801a-4028-a003-47fbe5190158","Type":"ContainerDied","Data":"4daecfe5e5fedc235149421381db7b1c6490e1e7548dbe656f57f2e1338c946e"} Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321634 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321676 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321742 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-config\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321775 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321829 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkjw2\" (UniqueName: \"kubernetes.io/projected/d9f8446d-e7a4-46b7-b62e-f4761a396473-kube-api-access-gkjw2\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.321856 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.322795 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.322832 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.323419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.323613 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.324444 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.324494 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9f8446d-e7a4-46b7-b62e-f4761a396473-config\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.355916 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkjw2\" (UniqueName: \"kubernetes.io/projected/d9f8446d-e7a4-46b7-b62e-f4761a396473-kube-api-access-gkjw2\") pod \"dnsmasq-dns-cb6ffcf87-clzsn\" (UID: \"d9f8446d-e7a4-46b7-b62e-f4761a396473\") " pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.436239 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.455360 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.526700 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.526796 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.526895 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.526958 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.527039 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn7ng\" (UniqueName: \"kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.527090 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb\") pod \"ef9ff948-801a-4028-a003-47fbe5190158\" (UID: \"ef9ff948-801a-4028-a003-47fbe5190158\") " Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.535253 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng" (OuterVolumeSpecName: "kube-api-access-rn7ng") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "kube-api-access-rn7ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.586577 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config" (OuterVolumeSpecName: "config") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.610450 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.615625 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.620082 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.626563 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef9ff948-801a-4028-a003-47fbe5190158" (UID: "ef9ff948-801a-4028-a003-47fbe5190158"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.642973 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.642990 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.643001 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.643010 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn7ng\" (UniqueName: \"kubernetes.io/projected/ef9ff948-801a-4028-a003-47fbe5190158-kube-api-access-rn7ng\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.643018 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.643026 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef9ff948-801a-4028-a003-47fbe5190158-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:27 crc kubenswrapper[4678]: I1013 13:03:27.917250 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-clzsn"] Oct 13 13:03:27 crc kubenswrapper[4678]: W1013 13:03:27.927500 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9f8446d_e7a4_46b7_b62e_f4761a396473.slice/crio-0e9052ed97b06f9a6866902ae9ab56f525f0ce2c6f02eb20093689b45c81a48c WatchSource:0}: Error finding container 0e9052ed97b06f9a6866902ae9ab56f525f0ce2c6f02eb20093689b45c81a48c: Status 404 returned error can't find the container with id 0e9052ed97b06f9a6866902ae9ab56f525f0ce2c6f02eb20093689b45c81a48c Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.248151 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9f8446d-e7a4-46b7-b62e-f4761a396473" containerID="e3324ad96a52b1c5a6eb4c8d522c64150b52214053cce695c9b32ca9eba588d8" exitCode=0 Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.248440 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" event={"ID":"d9f8446d-e7a4-46b7-b62e-f4761a396473","Type":"ContainerDied","Data":"e3324ad96a52b1c5a6eb4c8d522c64150b52214053cce695c9b32ca9eba588d8"} Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.248466 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" event={"ID":"d9f8446d-e7a4-46b7-b62e-f4761a396473","Type":"ContainerStarted","Data":"0e9052ed97b06f9a6866902ae9ab56f525f0ce2c6f02eb20093689b45c81a48c"} Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.250658 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" event={"ID":"ef9ff948-801a-4028-a003-47fbe5190158","Type":"ContainerDied","Data":"daff84c8ba73af9bf9bebe678cf2d27abbab0c484cc4d47408e1340233f3a5d7"} Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.250715 4678 scope.go:117] "RemoveContainer" containerID="4daecfe5e5fedc235149421381db7b1c6490e1e7548dbe656f57f2e1338c946e" Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.250722 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-ssjrr" Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.452505 4678 scope.go:117] "RemoveContainer" containerID="83ff36b66a33aa4a0d195f0f9a5ec6c09ac3ad476cdeec458b9001ab52477501" Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.454754 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.465074 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-ssjrr"] Oct 13 13:03:28 crc kubenswrapper[4678]: I1013 13:03:28.603457 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9ff948-801a-4028-a003-47fbe5190158" path="/var/lib/kubelet/pods/ef9ff948-801a-4028-a003-47fbe5190158/volumes" Oct 13 13:03:29 crc kubenswrapper[4678]: I1013 13:03:29.265079 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" event={"ID":"d9f8446d-e7a4-46b7-b62e-f4761a396473","Type":"ContainerStarted","Data":"57d7b4a227ffda7efce56f9859ade89e65ff7aa4de69473528967f4cc5ef57bf"} Oct 13 13:03:29 crc kubenswrapper[4678]: I1013 13:03:29.265224 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:29 crc kubenswrapper[4678]: I1013 13:03:29.290321 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" podStartSLOduration=2.290298285 podStartE2EDuration="2.290298285s" podCreationTimestamp="2025-10-13 13:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:03:29.287637147 +0000 UTC m=+1197.372175051" watchObservedRunningTime="2025-10-13 13:03:29.290298285 +0000 UTC m=+1197.374836209" Oct 13 13:03:37 crc kubenswrapper[4678]: I1013 13:03:37.437282 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-clzsn" Oct 13 13:03:37 crc kubenswrapper[4678]: I1013 13:03:37.524875 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:37 crc kubenswrapper[4678]: I1013 13:03:37.525266 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="dnsmasq-dns" containerID="cri-o://458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe" gracePeriod=10 Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.072773 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165025 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165134 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165248 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165323 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8pz2\" (UniqueName: \"kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165392 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165479 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.165520 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb\") pod \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\" (UID: \"13e7b0b7-853b-4d57-b959-83c1bdc457a1\") " Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.182086 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2" (OuterVolumeSpecName: "kube-api-access-j8pz2") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "kube-api-access-j8pz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.226876 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.238444 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.240204 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.241595 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.249763 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.252641 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config" (OuterVolumeSpecName: "config") pod "13e7b0b7-853b-4d57-b959-83c1bdc457a1" (UID: "13e7b0b7-853b-4d57-b959-83c1bdc457a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267587 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267618 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8pz2\" (UniqueName: \"kubernetes.io/projected/13e7b0b7-853b-4d57-b959-83c1bdc457a1-kube-api-access-j8pz2\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267630 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267639 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267649 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267657 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.267665 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7b0b7-853b-4d57-b959-83c1bdc457a1-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.367083 4678 generic.go:334] "Generic (PLEG): container finished" podID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerID="458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe" exitCode=0 Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.367123 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" event={"ID":"13e7b0b7-853b-4d57-b959-83c1bdc457a1","Type":"ContainerDied","Data":"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe"} Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.367147 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" event={"ID":"13e7b0b7-853b-4d57-b959-83c1bdc457a1","Type":"ContainerDied","Data":"27b51971375fa752335bde328aa933873e53a00bbcb14a972b46c27b4fdb91fc"} Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.367163 4678 scope.go:117] "RemoveContainer" containerID="458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.367278 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-bzkqx" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.393289 4678 scope.go:117] "RemoveContainer" containerID="e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.431655 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.434920 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-bzkqx"] Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.443461 4678 scope.go:117] "RemoveContainer" containerID="458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe" Oct 13 13:03:38 crc kubenswrapper[4678]: E1013 13:03:38.447164 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe\": container with ID starting with 458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe not found: ID does not exist" containerID="458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.447205 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe"} err="failed to get container status \"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe\": rpc error: code = NotFound desc = could not find container \"458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe\": container with ID starting with 458a9385a5d14055bedeaf89b037e250ed8bca7832ed0ea160d9b69e2446e9fe not found: ID does not exist" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.447231 4678 scope.go:117] "RemoveContainer" containerID="e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479" Oct 13 13:03:38 crc kubenswrapper[4678]: E1013 13:03:38.451143 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479\": container with ID starting with e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479 not found: ID does not exist" containerID="e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.451170 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479"} err="failed to get container status \"e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479\": rpc error: code = NotFound desc = could not find container \"e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479\": container with ID starting with e88f03e5dbe40951c8834a0c37b7fa186f8b275a67965f65d9e262092dcc3479 not found: ID does not exist" Oct 13 13:03:38 crc kubenswrapper[4678]: I1013 13:03:38.603930 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" path="/var/lib/kubelet/pods/13e7b0b7-853b-4d57-b959-83c1bdc457a1/volumes" Oct 13 13:03:50 crc kubenswrapper[4678]: I1013 13:03:50.519565 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c44f667-45a9-4e57-b8f9-ffd8da960a6e" containerID="0fb71cf7c9a1965b38ab34659773667d1c072c73b601b18a3c98dbae3096de20" exitCode=0 Oct 13 13:03:50 crc kubenswrapper[4678]: I1013 13:03:50.519642 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1c44f667-45a9-4e57-b8f9-ffd8da960a6e","Type":"ContainerDied","Data":"0fb71cf7c9a1965b38ab34659773667d1c072c73b601b18a3c98dbae3096de20"} Oct 13 13:03:50 crc kubenswrapper[4678]: I1013 13:03:50.523392 4678 generic.go:334] "Generic (PLEG): container finished" podID="a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe" containerID="e8bacb69b0f4c84029a274231e9d462b7cd53a43bb56b52efbf18b83c587f533" exitCode=0 Oct 13 13:03:50 crc kubenswrapper[4678]: I1013 13:03:50.523435 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe","Type":"ContainerDied","Data":"e8bacb69b0f4c84029a274231e9d462b7cd53a43bb56b52efbf18b83c587f533"} Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.021116 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw"] Oct 13 13:03:51 crc kubenswrapper[4678]: E1013 13:03:51.064586 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="init" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.064647 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="init" Oct 13 13:03:51 crc kubenswrapper[4678]: E1013 13:03:51.064711 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.064720 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: E1013 13:03:51.064741 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="init" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.064747 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="init" Oct 13 13:03:51 crc kubenswrapper[4678]: E1013 13:03:51.064759 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.064769 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.065453 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e7b0b7-853b-4d57-b959-83c1bdc457a1" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.065484 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9ff948-801a-4028-a003-47fbe5190158" containerName="dnsmasq-dns" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.066723 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw"] Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.066875 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.070382 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.070572 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.071214 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.071340 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.159236 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.159335 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hffd\" (UniqueName: \"kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.159369 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.159480 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.261199 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.261309 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.261366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hffd\" (UniqueName: \"kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.261408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.265870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.266170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.266338 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.286767 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hffd\" (UniqueName: \"kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.399197 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.536926 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe","Type":"ContainerStarted","Data":"e181fbd4318493f1638d766951edd6eb539ffe979e5b146f2ca8007bb85bc9ef"} Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.537139 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.539045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"1c44f667-45a9-4e57-b8f9-ffd8da960a6e","Type":"ContainerStarted","Data":"07f8c3e35a3abc0fe2bf64fcbf687db9a33f324ce5bcb0b0b078e30f8b054655"} Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.539313 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.577644 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.57762394 podStartE2EDuration="36.57762394s" podCreationTimestamp="2025-10-13 13:03:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:03:51.567793667 +0000 UTC m=+1219.652331551" watchObservedRunningTime="2025-10-13 13:03:51.57762394 +0000 UTC m=+1219.662161824" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.621175 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.62115795 podStartE2EDuration="37.62115795s" podCreationTimestamp="2025-10-13 13:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:03:51.6114457 +0000 UTC m=+1219.695983594" watchObservedRunningTime="2025-10-13 13:03:51.62115795 +0000 UTC m=+1219.705695834" Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.967333 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw"] Oct 13 13:03:51 crc kubenswrapper[4678]: W1013 13:03:51.978752 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4c81445_8564_47a3_ab04_db591cbf1a86.slice/crio-529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5 WatchSource:0}: Error finding container 529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5: Status 404 returned error can't find the container with id 529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5 Oct 13 13:03:51 crc kubenswrapper[4678]: I1013 13:03:51.981900 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:03:52 crc kubenswrapper[4678]: I1013 13:03:52.554233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" event={"ID":"d4c81445-8564-47a3-ab04-db591cbf1a86","Type":"ContainerStarted","Data":"529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5"} Oct 13 13:04:04 crc kubenswrapper[4678]: I1013 13:04:04.719206 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 13 13:04:05 crc kubenswrapper[4678]: I1013 13:04:05.503348 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 13 13:04:05 crc kubenswrapper[4678]: I1013 13:04:05.688529 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" event={"ID":"d4c81445-8564-47a3-ab04-db591cbf1a86","Type":"ContainerStarted","Data":"500b7c7d8a4e36c12aebf18a2ddf6f7b88c11b993a6deb02d1df868b9e491ddc"} Oct 13 13:04:05 crc kubenswrapper[4678]: I1013 13:04:05.713108 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" podStartSLOduration=3.252844629 podStartE2EDuration="15.713083156s" podCreationTimestamp="2025-10-13 13:03:50 +0000 UTC" firstStartedPulling="2025-10-13 13:03:51.981708354 +0000 UTC m=+1220.066246238" lastFinishedPulling="2025-10-13 13:04:04.441946891 +0000 UTC m=+1232.526484765" observedRunningTime="2025-10-13 13:04:05.707695888 +0000 UTC m=+1233.792233782" watchObservedRunningTime="2025-10-13 13:04:05.713083156 +0000 UTC m=+1233.797621050" Oct 13 13:04:16 crc kubenswrapper[4678]: I1013 13:04:16.831716 4678 generic.go:334] "Generic (PLEG): container finished" podID="d4c81445-8564-47a3-ab04-db591cbf1a86" containerID="500b7c7d8a4e36c12aebf18a2ddf6f7b88c11b993a6deb02d1df868b9e491ddc" exitCode=0 Oct 13 13:04:16 crc kubenswrapper[4678]: I1013 13:04:16.831806 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" event={"ID":"d4c81445-8564-47a3-ab04-db591cbf1a86","Type":"ContainerDied","Data":"500b7c7d8a4e36c12aebf18a2ddf6f7b88c11b993a6deb02d1df868b9e491ddc"} Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.334965 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.510905 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hffd\" (UniqueName: \"kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd\") pod \"d4c81445-8564-47a3-ab04-db591cbf1a86\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.511182 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory\") pod \"d4c81445-8564-47a3-ab04-db591cbf1a86\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.511308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key\") pod \"d4c81445-8564-47a3-ab04-db591cbf1a86\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.511371 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle\") pod \"d4c81445-8564-47a3-ab04-db591cbf1a86\" (UID: \"d4c81445-8564-47a3-ab04-db591cbf1a86\") " Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.519765 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd" (OuterVolumeSpecName: "kube-api-access-6hffd") pod "d4c81445-8564-47a3-ab04-db591cbf1a86" (UID: "d4c81445-8564-47a3-ab04-db591cbf1a86"). InnerVolumeSpecName "kube-api-access-6hffd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.519924 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d4c81445-8564-47a3-ab04-db591cbf1a86" (UID: "d4c81445-8564-47a3-ab04-db591cbf1a86"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.556880 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory" (OuterVolumeSpecName: "inventory") pod "d4c81445-8564-47a3-ab04-db591cbf1a86" (UID: "d4c81445-8564-47a3-ab04-db591cbf1a86"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.557590 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4c81445-8564-47a3-ab04-db591cbf1a86" (UID: "d4c81445-8564-47a3-ab04-db591cbf1a86"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.614579 4678 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.614632 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hffd\" (UniqueName: \"kubernetes.io/projected/d4c81445-8564-47a3-ab04-db591cbf1a86-kube-api-access-6hffd\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.614655 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.614674 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4c81445-8564-47a3-ab04-db591cbf1a86-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.864118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" event={"ID":"d4c81445-8564-47a3-ab04-db591cbf1a86","Type":"ContainerDied","Data":"529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5"} Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.864162 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="529064f73774c6b291fd25ead4daa7da2af191db9baff457464b12e36a4247b5" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.864220 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.974509 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt"] Oct 13 13:04:18 crc kubenswrapper[4678]: E1013 13:04:18.975339 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4c81445-8564-47a3-ab04-db591cbf1a86" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.975368 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4c81445-8564-47a3-ab04-db591cbf1a86" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.976575 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4c81445-8564-47a3-ab04-db591cbf1a86" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.977360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.981299 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.983817 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.984128 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.986308 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:04:18 crc kubenswrapper[4678]: I1013 13:04:18.995478 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt"] Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.124116 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj2zq\" (UniqueName: \"kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.124180 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.124225 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.226137 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj2zq\" (UniqueName: \"kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.226244 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.226304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.232233 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.232397 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.261497 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj2zq\" (UniqueName: \"kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-s9vpt\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.302673 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.838458 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt"] Oct 13 13:04:19 crc kubenswrapper[4678]: I1013 13:04:19.874984 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" event={"ID":"26b2ddaa-5eaa-48c5-947b-507b97253a67","Type":"ContainerStarted","Data":"7d43a54cd776b389811e257175354b1f8a5e0366adcc515d239063f5bf971ba8"} Oct 13 13:04:20 crc kubenswrapper[4678]: I1013 13:04:20.892319 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" event={"ID":"26b2ddaa-5eaa-48c5-947b-507b97253a67","Type":"ContainerStarted","Data":"8499b6c335a8d8fb13063ac8b78b9a871f4f100a030467a7974375b2587755b1"} Oct 13 13:04:20 crc kubenswrapper[4678]: I1013 13:04:20.927909 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" podStartSLOduration=2.416526592 podStartE2EDuration="2.927879915s" podCreationTimestamp="2025-10-13 13:04:18 +0000 UTC" firstStartedPulling="2025-10-13 13:04:19.842823665 +0000 UTC m=+1247.927361559" lastFinishedPulling="2025-10-13 13:04:20.354176968 +0000 UTC m=+1248.438714882" observedRunningTime="2025-10-13 13:04:20.912332155 +0000 UTC m=+1248.996870099" watchObservedRunningTime="2025-10-13 13:04:20.927879915 +0000 UTC m=+1249.012417839" Oct 13 13:04:23 crc kubenswrapper[4678]: I1013 13:04:23.932023 4678 generic.go:334] "Generic (PLEG): container finished" podID="26b2ddaa-5eaa-48c5-947b-507b97253a67" containerID="8499b6c335a8d8fb13063ac8b78b9a871f4f100a030467a7974375b2587755b1" exitCode=0 Oct 13 13:04:23 crc kubenswrapper[4678]: I1013 13:04:23.932088 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" event={"ID":"26b2ddaa-5eaa-48c5-947b-507b97253a67","Type":"ContainerDied","Data":"8499b6c335a8d8fb13063ac8b78b9a871f4f100a030467a7974375b2587755b1"} Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.425749 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.506366 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.506417 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.560592 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key\") pod \"26b2ddaa-5eaa-48c5-947b-507b97253a67\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.560632 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory\") pod \"26b2ddaa-5eaa-48c5-947b-507b97253a67\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.560764 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj2zq\" (UniqueName: \"kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq\") pod \"26b2ddaa-5eaa-48c5-947b-507b97253a67\" (UID: \"26b2ddaa-5eaa-48c5-947b-507b97253a67\") " Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.567388 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq" (OuterVolumeSpecName: "kube-api-access-sj2zq") pod "26b2ddaa-5eaa-48c5-947b-507b97253a67" (UID: "26b2ddaa-5eaa-48c5-947b-507b97253a67"). InnerVolumeSpecName "kube-api-access-sj2zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.608741 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory" (OuterVolumeSpecName: "inventory") pod "26b2ddaa-5eaa-48c5-947b-507b97253a67" (UID: "26b2ddaa-5eaa-48c5-947b-507b97253a67"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.614218 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "26b2ddaa-5eaa-48c5-947b-507b97253a67" (UID: "26b2ddaa-5eaa-48c5-947b-507b97253a67"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.663197 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj2zq\" (UniqueName: \"kubernetes.io/projected/26b2ddaa-5eaa-48c5-947b-507b97253a67-kube-api-access-sj2zq\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.663227 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.663240 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/26b2ddaa-5eaa-48c5-947b-507b97253a67-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.958625 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.958684 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-s9vpt" event={"ID":"26b2ddaa-5eaa-48c5-947b-507b97253a67","Type":"ContainerDied","Data":"7d43a54cd776b389811e257175354b1f8a5e0366adcc515d239063f5bf971ba8"} Oct 13 13:04:25 crc kubenswrapper[4678]: I1013 13:04:25.958729 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d43a54cd776b389811e257175354b1f8a5e0366adcc515d239063f5bf971ba8" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.040014 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf"] Oct 13 13:04:26 crc kubenswrapper[4678]: E1013 13:04:26.040488 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b2ddaa-5eaa-48c5-947b-507b97253a67" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.040511 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b2ddaa-5eaa-48c5-947b-507b97253a67" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.040776 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b2ddaa-5eaa-48c5-947b-507b97253a67" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.041539 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.046211 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.049490 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.049958 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.051300 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.062906 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf"] Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.173934 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.174201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.174604 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.174639 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhdk5\" (UniqueName: \"kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.276907 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.276975 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhdk5\" (UniqueName: \"kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.277096 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.277213 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.282893 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.288898 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.289244 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.297417 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhdk5\" (UniqueName: \"kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:26 crc kubenswrapper[4678]: I1013 13:04:26.415544 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:04:27 crc kubenswrapper[4678]: I1013 13:04:27.017647 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf"] Oct 13 13:04:27 crc kubenswrapper[4678]: W1013 13:04:27.028596 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31ca8ec6_826e_4f5b_9e58_160e18674d9f.slice/crio-ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6 WatchSource:0}: Error finding container ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6: Status 404 returned error can't find the container with id ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6 Oct 13 13:04:27 crc kubenswrapper[4678]: I1013 13:04:27.986213 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" event={"ID":"31ca8ec6-826e-4f5b-9e58-160e18674d9f","Type":"ContainerStarted","Data":"59c588d6efeac27e96ee9c9a04320d450bce00e733b53fc8b6c7668bfa54f15a"} Oct 13 13:04:27 crc kubenswrapper[4678]: I1013 13:04:27.986603 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" event={"ID":"31ca8ec6-826e-4f5b-9e58-160e18674d9f","Type":"ContainerStarted","Data":"ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6"} Oct 13 13:04:28 crc kubenswrapper[4678]: I1013 13:04:28.007752 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" podStartSLOduration=1.374161363 podStartE2EDuration="2.00773597s" podCreationTimestamp="2025-10-13 13:04:26 +0000 UTC" firstStartedPulling="2025-10-13 13:04:27.033021019 +0000 UTC m=+1255.117558903" lastFinishedPulling="2025-10-13 13:04:27.666595616 +0000 UTC m=+1255.751133510" observedRunningTime="2025-10-13 13:04:28.004394325 +0000 UTC m=+1256.088932219" watchObservedRunningTime="2025-10-13 13:04:28.00773597 +0000 UTC m=+1256.092273854" Oct 13 13:04:52 crc kubenswrapper[4678]: I1013 13:04:52.717256 4678 scope.go:117] "RemoveContainer" containerID="2e7cdfe2be6d3789715e07cdb62d07d9714fabf3cfb95e224b3b886b7c1e720f" Oct 13 13:04:55 crc kubenswrapper[4678]: I1013 13:04:55.505920 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:04:55 crc kubenswrapper[4678]: I1013 13:04:55.506606 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.506729 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.507293 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.507343 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.508141 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.508204 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7" gracePeriod=600 Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.646085 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7" exitCode=0 Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.646137 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7"} Oct 13 13:05:25 crc kubenswrapper[4678]: I1013 13:05:25.646171 4678 scope.go:117] "RemoveContainer" containerID="4e2dabd2a11cf6969bd896a14eb48fa4b150531a592ec3c31cb880ed3f35cb06" Oct 13 13:05:26 crc kubenswrapper[4678]: I1013 13:05:26.659905 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98"} Oct 13 13:05:52 crc kubenswrapper[4678]: I1013 13:05:52.819871 4678 scope.go:117] "RemoveContainer" containerID="dfd73890b80623aa90907c6d3ab182b3541e6786fdd780298301bd2720a93d6d" Oct 13 13:05:52 crc kubenswrapper[4678]: I1013 13:05:52.851494 4678 scope.go:117] "RemoveContainer" containerID="0f8368914f3f4ee2fb0fd6444e6eeaf5be68327054789154ac03c48228a15c63" Oct 13 13:05:52 crc kubenswrapper[4678]: I1013 13:05:52.873792 4678 scope.go:117] "RemoveContainer" containerID="ed14b44f1295610b341995a2d0a87e6a0584d0632b5b219cf3996b516f921889" Oct 13 13:05:52 crc kubenswrapper[4678]: I1013 13:05:52.923875 4678 scope.go:117] "RemoveContainer" containerID="0f6deaacf2953924ac56461237f19f9d5d0f1a8c394a76f29e0405508a446019" Oct 13 13:07:25 crc kubenswrapper[4678]: I1013 13:07:25.505801 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:07:25 crc kubenswrapper[4678]: I1013 13:07:25.506412 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:07:35 crc kubenswrapper[4678]: I1013 13:07:35.026661 4678 generic.go:334] "Generic (PLEG): container finished" podID="31ca8ec6-826e-4f5b-9e58-160e18674d9f" containerID="59c588d6efeac27e96ee9c9a04320d450bce00e733b53fc8b6c7668bfa54f15a" exitCode=0 Oct 13 13:07:35 crc kubenswrapper[4678]: I1013 13:07:35.026729 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" event={"ID":"31ca8ec6-826e-4f5b-9e58-160e18674d9f","Type":"ContainerDied","Data":"59c588d6efeac27e96ee9c9a04320d450bce00e733b53fc8b6c7668bfa54f15a"} Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.431086 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.512886 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhdk5\" (UniqueName: \"kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5\") pod \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.512956 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle\") pod \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.513045 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key\") pod \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.513227 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory\") pod \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\" (UID: \"31ca8ec6-826e-4f5b-9e58-160e18674d9f\") " Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.520209 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5" (OuterVolumeSpecName: "kube-api-access-rhdk5") pod "31ca8ec6-826e-4f5b-9e58-160e18674d9f" (UID: "31ca8ec6-826e-4f5b-9e58-160e18674d9f"). InnerVolumeSpecName "kube-api-access-rhdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.522454 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "31ca8ec6-826e-4f5b-9e58-160e18674d9f" (UID: "31ca8ec6-826e-4f5b-9e58-160e18674d9f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.545794 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory" (OuterVolumeSpecName: "inventory") pod "31ca8ec6-826e-4f5b-9e58-160e18674d9f" (UID: "31ca8ec6-826e-4f5b-9e58-160e18674d9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.551029 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "31ca8ec6-826e-4f5b-9e58-160e18674d9f" (UID: "31ca8ec6-826e-4f5b-9e58-160e18674d9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.615575 4678 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.615611 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.615620 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/31ca8ec6-826e-4f5b-9e58-160e18674d9f-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:07:36 crc kubenswrapper[4678]: I1013 13:07:36.615628 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhdk5\" (UniqueName: \"kubernetes.io/projected/31ca8ec6-826e-4f5b-9e58-160e18674d9f-kube-api-access-rhdk5\") on node \"crc\" DevicePath \"\"" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.044728 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" event={"ID":"31ca8ec6-826e-4f5b-9e58-160e18674d9f","Type":"ContainerDied","Data":"ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6"} Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.044772 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffb8fce9b91747ac402f852c44b64fcd5d8439a4183e15383316c3833d722df6" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.044785 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.151955 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894"] Oct 13 13:07:37 crc kubenswrapper[4678]: E1013 13:07:37.152482 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ca8ec6-826e-4f5b-9e58-160e18674d9f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.152507 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ca8ec6-826e-4f5b-9e58-160e18674d9f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.152776 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ca8ec6-826e-4f5b-9e58-160e18674d9f" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.153595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.156119 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.157111 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.157161 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.157742 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.161789 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894"] Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.240173 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.240253 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.240374 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7ts7\" (UniqueName: \"kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.342245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.342316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.342424 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7ts7\" (UniqueName: \"kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.345511 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.347801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.357641 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7ts7\" (UniqueName: \"kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hn894\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:37 crc kubenswrapper[4678]: I1013 13:07:37.468788 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:07:38 crc kubenswrapper[4678]: I1013 13:07:38.133741 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894"] Oct 13 13:07:39 crc kubenswrapper[4678]: I1013 13:07:39.081514 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" event={"ID":"48c09433-04bb-4cd0-8bfd-72606ebb285a","Type":"ContainerStarted","Data":"d8289f6f629becc1f52dc2348102dc199c780a7399325fed4919921d91bfc113"} Oct 13 13:07:39 crc kubenswrapper[4678]: I1013 13:07:39.081867 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" event={"ID":"48c09433-04bb-4cd0-8bfd-72606ebb285a","Type":"ContainerStarted","Data":"da589256880105ea57a0f362ae9b47eb0068a5d0989b9cef75ee08e2f3632255"} Oct 13 13:07:39 crc kubenswrapper[4678]: I1013 13:07:39.099999 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" podStartSLOduration=1.70296732 podStartE2EDuration="2.099978494s" podCreationTimestamp="2025-10-13 13:07:37 +0000 UTC" firstStartedPulling="2025-10-13 13:07:38.141014755 +0000 UTC m=+1446.225552649" lastFinishedPulling="2025-10-13 13:07:38.538025939 +0000 UTC m=+1446.622563823" observedRunningTime="2025-10-13 13:07:39.098118086 +0000 UTC m=+1447.182655990" watchObservedRunningTime="2025-10-13 13:07:39.099978494 +0000 UTC m=+1447.184516378" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.171431 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.173870 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.226926 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.252252 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.252401 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9l5z\" (UniqueName: \"kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.252438 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.353784 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9l5z\" (UniqueName: \"kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.354112 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.354308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.354716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.354866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.383109 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9l5z\" (UniqueName: \"kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z\") pod \"community-operators-mbhjx\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:48 crc kubenswrapper[4678]: I1013 13:07:48.490764 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:49 crc kubenswrapper[4678]: I1013 13:07:49.056527 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:07:49 crc kubenswrapper[4678]: I1013 13:07:49.193293 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerStarted","Data":"330470ea3d87d805dc128937a5e885a2dc933d91d640e39a61ae99e127a506b5"} Oct 13 13:07:50 crc kubenswrapper[4678]: I1013 13:07:50.202296 4678 generic.go:334] "Generic (PLEG): container finished" podID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerID="1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800" exitCode=0 Oct 13 13:07:50 crc kubenswrapper[4678]: I1013 13:07:50.202394 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerDied","Data":"1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800"} Oct 13 13:07:51 crc kubenswrapper[4678]: I1013 13:07:51.213848 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerStarted","Data":"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3"} Oct 13 13:07:52 crc kubenswrapper[4678]: I1013 13:07:52.226541 4678 generic.go:334] "Generic (PLEG): container finished" podID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerID="853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3" exitCode=0 Oct 13 13:07:52 crc kubenswrapper[4678]: I1013 13:07:52.226632 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerDied","Data":"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3"} Oct 13 13:07:53 crc kubenswrapper[4678]: I1013 13:07:53.078456 4678 scope.go:117] "RemoveContainer" containerID="67cee14a7b9c10957db18a2ced5a33dbc9c9f862cce7b649da2cc5958faeaa00" Oct 13 13:07:53 crc kubenswrapper[4678]: I1013 13:07:53.127630 4678 scope.go:117] "RemoveContainer" containerID="32a1c32c9360782e158ab6e81d9fc444a916a8e3a5e1c97580dd75717a0835ec" Oct 13 13:07:53 crc kubenswrapper[4678]: I1013 13:07:53.241328 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerStarted","Data":"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c"} Oct 13 13:07:53 crc kubenswrapper[4678]: I1013 13:07:53.259342 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mbhjx" podStartSLOduration=2.711198855 podStartE2EDuration="5.259325768s" podCreationTimestamp="2025-10-13 13:07:48 +0000 UTC" firstStartedPulling="2025-10-13 13:07:50.204078603 +0000 UTC m=+1458.288616487" lastFinishedPulling="2025-10-13 13:07:52.752205516 +0000 UTC m=+1460.836743400" observedRunningTime="2025-10-13 13:07:53.257596683 +0000 UTC m=+1461.342134577" watchObservedRunningTime="2025-10-13 13:07:53.259325768 +0000 UTC m=+1461.343863652" Oct 13 13:07:55 crc kubenswrapper[4678]: I1013 13:07:55.505883 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:07:55 crc kubenswrapper[4678]: I1013 13:07:55.506288 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:07:58 crc kubenswrapper[4678]: I1013 13:07:58.491932 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:58 crc kubenswrapper[4678]: I1013 13:07:58.492165 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:58 crc kubenswrapper[4678]: I1013 13:07:58.585245 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:59 crc kubenswrapper[4678]: I1013 13:07:59.379551 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:07:59 crc kubenswrapper[4678]: I1013 13:07:59.440346 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.335354 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mbhjx" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="registry-server" containerID="cri-o://aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c" gracePeriod=2 Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.808007 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.980146 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content\") pod \"28bfcb0d-7478-45ed-a1c8-fb306d924100\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.980328 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9l5z\" (UniqueName: \"kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z\") pod \"28bfcb0d-7478-45ed-a1c8-fb306d924100\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.981104 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities" (OuterVolumeSpecName: "utilities") pod "28bfcb0d-7478-45ed-a1c8-fb306d924100" (UID: "28bfcb0d-7478-45ed-a1c8-fb306d924100"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.981165 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities\") pod \"28bfcb0d-7478-45ed-a1c8-fb306d924100\" (UID: \"28bfcb0d-7478-45ed-a1c8-fb306d924100\") " Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.981673 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:01 crc kubenswrapper[4678]: I1013 13:08:01.997700 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z" (OuterVolumeSpecName: "kube-api-access-r9l5z") pod "28bfcb0d-7478-45ed-a1c8-fb306d924100" (UID: "28bfcb0d-7478-45ed-a1c8-fb306d924100"). InnerVolumeSpecName "kube-api-access-r9l5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.019630 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28bfcb0d-7478-45ed-a1c8-fb306d924100" (UID: "28bfcb0d-7478-45ed-a1c8-fb306d924100"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.083607 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28bfcb0d-7478-45ed-a1c8-fb306d924100-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.083642 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9l5z\" (UniqueName: \"kubernetes.io/projected/28bfcb0d-7478-45ed-a1c8-fb306d924100-kube-api-access-r9l5z\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.347786 4678 generic.go:334] "Generic (PLEG): container finished" podID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerID="aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c" exitCode=0 Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.347826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerDied","Data":"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c"} Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.347894 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mbhjx" event={"ID":"28bfcb0d-7478-45ed-a1c8-fb306d924100","Type":"ContainerDied","Data":"330470ea3d87d805dc128937a5e885a2dc933d91d640e39a61ae99e127a506b5"} Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.347922 4678 scope.go:117] "RemoveContainer" containerID="aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.347932 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mbhjx" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.372816 4678 scope.go:117] "RemoveContainer" containerID="853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.393769 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.400394 4678 scope.go:117] "RemoveContainer" containerID="1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.405134 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mbhjx"] Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.458031 4678 scope.go:117] "RemoveContainer" containerID="aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c" Oct 13 13:08:02 crc kubenswrapper[4678]: E1013 13:08:02.458425 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c\": container with ID starting with aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c not found: ID does not exist" containerID="aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.458462 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c"} err="failed to get container status \"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c\": rpc error: code = NotFound desc = could not find container \"aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c\": container with ID starting with aea6f7b6bef74810f911ad56f87b25a840e16d3675e94dbd53f74e5a2527518c not found: ID does not exist" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.458483 4678 scope.go:117] "RemoveContainer" containerID="853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3" Oct 13 13:08:02 crc kubenswrapper[4678]: E1013 13:08:02.458700 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3\": container with ID starting with 853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3 not found: ID does not exist" containerID="853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.458723 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3"} err="failed to get container status \"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3\": rpc error: code = NotFound desc = could not find container \"853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3\": container with ID starting with 853a1466dc3ece4b8cc762053b44ff5573548439183ed94b432ec771d21c2ad3 not found: ID does not exist" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.458743 4678 scope.go:117] "RemoveContainer" containerID="1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800" Oct 13 13:08:02 crc kubenswrapper[4678]: E1013 13:08:02.459012 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800\": container with ID starting with 1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800 not found: ID does not exist" containerID="1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.459037 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800"} err="failed to get container status \"1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800\": rpc error: code = NotFound desc = could not find container \"1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800\": container with ID starting with 1c4d3a59ff70e04d137d85cbf3646c379d0f49caf4b8c9101e63443bc074e800 not found: ID does not exist" Oct 13 13:08:02 crc kubenswrapper[4678]: I1013 13:08:02.604935 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" path="/var/lib/kubelet/pods/28bfcb0d-7478-45ed-a1c8-fb306d924100/volumes" Oct 13 13:08:25 crc kubenswrapper[4678]: I1013 13:08:25.506518 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:08:25 crc kubenswrapper[4678]: I1013 13:08:25.507032 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:08:25 crc kubenswrapper[4678]: I1013 13:08:25.507083 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:08:25 crc kubenswrapper[4678]: I1013 13:08:25.507745 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:08:25 crc kubenswrapper[4678]: I1013 13:08:25.507797 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" gracePeriod=600 Oct 13 13:08:25 crc kubenswrapper[4678]: E1013 13:08:25.631103 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:08:26 crc kubenswrapper[4678]: I1013 13:08:26.626226 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" exitCode=0 Oct 13 13:08:26 crc kubenswrapper[4678]: I1013 13:08:26.626433 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98"} Oct 13 13:08:26 crc kubenswrapper[4678]: I1013 13:08:26.626604 4678 scope.go:117] "RemoveContainer" containerID="fb46f44b44a86aa2b285410743355f9772e39c34f1ba5bdbefde398832248ab7" Oct 13 13:08:26 crc kubenswrapper[4678]: I1013 13:08:26.627219 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:08:26 crc kubenswrapper[4678]: E1013 13:08:26.627457 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.057385 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-d4hbz"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.072048 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-dv4hj"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.080817 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vgs9f"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.087992 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-d4hbz"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.096506 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vgs9f"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.103637 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-dv4hj"] Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.608699 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159" path="/var/lib/kubelet/pods/3fae8bdb-cfa7-4d91-bd3c-ddd8efc2d159/volumes" Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.609435 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80039354-df5b-44f1-a7c5-1b997e49229a" path="/var/lib/kubelet/pods/80039354-df5b-44f1-a7c5-1b997e49229a/volumes" Oct 13 13:08:30 crc kubenswrapper[4678]: I1013 13:08:30.610132 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac932bc-64cc-4681-b720-dc6f2ee45f6c" path="/var/lib/kubelet/pods/9ac932bc-64cc-4681-b720-dc6f2ee45f6c/volumes" Oct 13 13:08:38 crc kubenswrapper[4678]: I1013 13:08:38.593114 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:08:38 crc kubenswrapper[4678]: E1013 13:08:38.594200 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.902145 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:39 crc kubenswrapper[4678]: E1013 13:08:39.902553 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="extract-utilities" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.902565 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="extract-utilities" Oct 13 13:08:39 crc kubenswrapper[4678]: E1013 13:08:39.902582 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="registry-server" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.902589 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="registry-server" Oct 13 13:08:39 crc kubenswrapper[4678]: E1013 13:08:39.902624 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="extract-content" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.902630 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="extract-content" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.902816 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="28bfcb0d-7478-45ed-a1c8-fb306d924100" containerName="registry-server" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.904305 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:39 crc kubenswrapper[4678]: I1013 13:08:39.921785 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.015884 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.015946 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.016261 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhxvd\" (UniqueName: \"kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.029098 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9be8-account-create-t74kb"] Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.041348 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9be8-account-create-t74kb"] Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.118076 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhxvd\" (UniqueName: \"kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.118421 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.118449 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.118869 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.119425 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.138247 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhxvd\" (UniqueName: \"kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd\") pod \"certified-operators-5rkml\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.226561 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.602671 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6d59036-e487-4f27-9522-4db47fe067f5" path="/var/lib/kubelet/pods/c6d59036-e487-4f27-9522-4db47fe067f5/volumes" Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.697258 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:40 crc kubenswrapper[4678]: W1013 13:08:40.700440 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68fb33a8_29dc_4468_9c4c_6b01e8d9a07e.slice/crio-1b460a35cc8876866d2b728054e1b5ac1680c8a7e75c4eea82618117fd184c6c WatchSource:0}: Error finding container 1b460a35cc8876866d2b728054e1b5ac1680c8a7e75c4eea82618117fd184c6c: Status 404 returned error can't find the container with id 1b460a35cc8876866d2b728054e1b5ac1680c8a7e75c4eea82618117fd184c6c Oct 13 13:08:40 crc kubenswrapper[4678]: I1013 13:08:40.772015 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerStarted","Data":"1b460a35cc8876866d2b728054e1b5ac1680c8a7e75c4eea82618117fd184c6c"} Oct 13 13:08:41 crc kubenswrapper[4678]: I1013 13:08:41.781649 4678 generic.go:334] "Generic (PLEG): container finished" podID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerID="5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891" exitCode=0 Oct 13 13:08:41 crc kubenswrapper[4678]: I1013 13:08:41.781943 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerDied","Data":"5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891"} Oct 13 13:08:42 crc kubenswrapper[4678]: I1013 13:08:42.794200 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerStarted","Data":"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754"} Oct 13 13:08:43 crc kubenswrapper[4678]: I1013 13:08:43.802794 4678 generic.go:334] "Generic (PLEG): container finished" podID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerID="6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754" exitCode=0 Oct 13 13:08:43 crc kubenswrapper[4678]: I1013 13:08:43.802855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerDied","Data":"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754"} Oct 13 13:08:44 crc kubenswrapper[4678]: I1013 13:08:44.029739 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5e8c-account-create-dhm4c"] Oct 13 13:08:44 crc kubenswrapper[4678]: I1013 13:08:44.036938 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5e8c-account-create-dhm4c"] Oct 13 13:08:44 crc kubenswrapper[4678]: I1013 13:08:44.613337 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d624a6f3-db83-4d66-b86a-0d76348a9839" path="/var/lib/kubelet/pods/d624a6f3-db83-4d66-b86a-0d76348a9839/volumes" Oct 13 13:08:44 crc kubenswrapper[4678]: I1013 13:08:44.812816 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerStarted","Data":"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c"} Oct 13 13:08:44 crc kubenswrapper[4678]: I1013 13:08:44.859210 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5rkml" podStartSLOduration=3.4070256309999998 podStartE2EDuration="5.859187185s" podCreationTimestamp="2025-10-13 13:08:39 +0000 UTC" firstStartedPulling="2025-10-13 13:08:41.783968991 +0000 UTC m=+1509.868506875" lastFinishedPulling="2025-10-13 13:08:44.236130545 +0000 UTC m=+1512.320668429" observedRunningTime="2025-10-13 13:08:44.830785667 +0000 UTC m=+1512.915323552" watchObservedRunningTime="2025-10-13 13:08:44.859187185 +0000 UTC m=+1512.943725099" Oct 13 13:08:45 crc kubenswrapper[4678]: I1013 13:08:45.033151 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-96f2-account-create-4h6m4"] Oct 13 13:08:45 crc kubenswrapper[4678]: I1013 13:08:45.040763 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-96f2-account-create-4h6m4"] Oct 13 13:08:46 crc kubenswrapper[4678]: I1013 13:08:46.613896 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efbce8cb-8fdf-4beb-b1c1-87937f5b7df4" path="/var/lib/kubelet/pods/efbce8cb-8fdf-4beb-b1c1-87937f5b7df4/volumes" Oct 13 13:08:50 crc kubenswrapper[4678]: I1013 13:08:50.227433 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:50 crc kubenswrapper[4678]: I1013 13:08:50.227810 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:50 crc kubenswrapper[4678]: I1013 13:08:50.306536 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:50 crc kubenswrapper[4678]: I1013 13:08:50.930120 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:51 crc kubenswrapper[4678]: I1013 13:08:51.003362 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:51 crc kubenswrapper[4678]: I1013 13:08:51.026805 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-7lkqp"] Oct 13 13:08:51 crc kubenswrapper[4678]: I1013 13:08:51.035291 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-7lkqp"] Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.022717 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-b62wr"] Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.031242 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-k5zwg"] Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.038640 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-k5zwg"] Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.045522 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-b62wr"] Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.600337 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:08:52 crc kubenswrapper[4678]: E1013 13:08:52.600669 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.601945 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6" path="/var/lib/kubelet/pods/1aa1f5d5-1a13-40bd-b4aa-5e5c602f73e6/volumes" Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.602721 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93fd1326-7d69-497f-a770-ee1fb9de923d" path="/var/lib/kubelet/pods/93fd1326-7d69-497f-a770-ee1fb9de923d/volumes" Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.603455 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdc002e-fb8c-4e54-8a8b-570e8b134cbc" path="/var/lib/kubelet/pods/efdc002e-fb8c-4e54-8a8b-570e8b134cbc/volumes" Oct 13 13:08:52 crc kubenswrapper[4678]: I1013 13:08:52.890568 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5rkml" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="registry-server" containerID="cri-o://e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c" gracePeriod=2 Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.184770 4678 scope.go:117] "RemoveContainer" containerID="701a66fd1f0a3896fed996332e51b2d169fc9d7b795b789a9deee5de1e1d7588" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.209391 4678 scope.go:117] "RemoveContainer" containerID="f938b2737eb47e152014cde3580556cad37961bdcea5df04ea7820970f20c5f5" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.364809 4678 scope.go:117] "RemoveContainer" containerID="05972cb67214c4e5f83b5ee0949d02e462d179944ab6bf6a92c7908c4b2e5ba1" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.376799 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.393936 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content\") pod \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.394173 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities\") pod \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.394261 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhxvd\" (UniqueName: \"kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd\") pod \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\" (UID: \"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e\") " Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.399701 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities" (OuterVolumeSpecName: "utilities") pod "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" (UID: "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.400259 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd" (OuterVolumeSpecName: "kube-api-access-rhxvd") pod "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" (UID: "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e"). InnerVolumeSpecName "kube-api-access-rhxvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.411993 4678 scope.go:117] "RemoveContainer" containerID="4224014be7090eb5eb60ade9164db368e982b89cc74f8221f1dc639f8ee9fc17" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.458870 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" (UID: "68fb33a8-29dc-4468-9c4c-6b01e8d9a07e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.459888 4678 scope.go:117] "RemoveContainer" containerID="dc471b38e88b76899004ebd4681428dfcb9b2ddb75edc5c151eb2c49ddd7ed06" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.484894 4678 scope.go:117] "RemoveContainer" containerID="30f18b2f11249fb9e8355d2a7261d818e09322dcccd4a1534cee5d3130b2b37a" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.496202 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.496233 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhxvd\" (UniqueName: \"kubernetes.io/projected/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-kube-api-access-rhxvd\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.496245 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.500856 4678 scope.go:117] "RemoveContainer" containerID="2932c5ee6c30cecce0d821a793803ace56c0cca8858436bbb85ba77df531bd83" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.518404 4678 scope.go:117] "RemoveContainer" containerID="56da00d5f38771b2ff822dc867eeb009b4feb7906cd960cfd1516c7ca70c7298" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.536965 4678 scope.go:117] "RemoveContainer" containerID="49232af032db8d67bcac68322ae85a833a7502f74add7088130fed6d4ce902c3" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.901689 4678 generic.go:334] "Generic (PLEG): container finished" podID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerID="e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c" exitCode=0 Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.901757 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerDied","Data":"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c"} Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.901768 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5rkml" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.902176 4678 scope.go:117] "RemoveContainer" containerID="e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.902084 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5rkml" event={"ID":"68fb33a8-29dc-4468-9c4c-6b01e8d9a07e","Type":"ContainerDied","Data":"1b460a35cc8876866d2b728054e1b5ac1680c8a7e75c4eea82618117fd184c6c"} Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.922537 4678 scope.go:117] "RemoveContainer" containerID="6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.950720 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.951028 4678 scope.go:117] "RemoveContainer" containerID="5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.960449 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5rkml"] Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.982099 4678 scope.go:117] "RemoveContainer" containerID="e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c" Oct 13 13:08:53 crc kubenswrapper[4678]: E1013 13:08:53.982580 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c\": container with ID starting with e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c not found: ID does not exist" containerID="e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.982648 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c"} err="failed to get container status \"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c\": rpc error: code = NotFound desc = could not find container \"e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c\": container with ID starting with e3ddde8207f4e037ca19c19f406b628bfd4193c9739e41155598e8b8754c548c not found: ID does not exist" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.982672 4678 scope.go:117] "RemoveContainer" containerID="6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754" Oct 13 13:08:53 crc kubenswrapper[4678]: E1013 13:08:53.983081 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754\": container with ID starting with 6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754 not found: ID does not exist" containerID="6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.983119 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754"} err="failed to get container status \"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754\": rpc error: code = NotFound desc = could not find container \"6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754\": container with ID starting with 6b69098a10be0028fb635c4d73ed58ff73144a0aeca6014d8426c5ea2520a754 not found: ID does not exist" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.983144 4678 scope.go:117] "RemoveContainer" containerID="5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891" Oct 13 13:08:53 crc kubenswrapper[4678]: E1013 13:08:53.983500 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891\": container with ID starting with 5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891 not found: ID does not exist" containerID="5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891" Oct 13 13:08:53 crc kubenswrapper[4678]: I1013 13:08:53.983638 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891"} err="failed to get container status \"5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891\": rpc error: code = NotFound desc = could not find container \"5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891\": container with ID starting with 5dcc3623a1b4e6e34d7910323a4bf2402927af95bf5cc4f29f8b21b45543f891 not found: ID does not exist" Oct 13 13:08:54 crc kubenswrapper[4678]: I1013 13:08:54.602645 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" path="/var/lib/kubelet/pods/68fb33a8-29dc-4468-9c4c-6b01e8d9a07e/volumes" Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.048479 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-3012-account-create-bhsxw"] Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.062556 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-06d9-account-create-tjsv6"] Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.075882 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-101c-account-create-n6vrn"] Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.085712 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-06d9-account-create-tjsv6"] Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.097234 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-3012-account-create-bhsxw"] Oct 13 13:09:05 crc kubenswrapper[4678]: I1013 13:09:05.107290 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-101c-account-create-n6vrn"] Oct 13 13:09:06 crc kubenswrapper[4678]: I1013 13:09:06.624493 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d1f30c5-e273-4bd7-9d62-62f26b253fca" path="/var/lib/kubelet/pods/0d1f30c5-e273-4bd7-9d62-62f26b253fca/volumes" Oct 13 13:09:06 crc kubenswrapper[4678]: I1013 13:09:06.626563 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dbff364-1cc3-44d1-af82-63bbbc39d2e3" path="/var/lib/kubelet/pods/6dbff364-1cc3-44d1-af82-63bbbc39d2e3/volumes" Oct 13 13:09:06 crc kubenswrapper[4678]: I1013 13:09:06.627377 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9182c1b1-e2f1-48b6-85ec-459d42ccc2ce" path="/var/lib/kubelet/pods/9182c1b1-e2f1-48b6-85ec-459d42ccc2ce/volumes" Oct 13 13:09:07 crc kubenswrapper[4678]: I1013 13:09:07.592679 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:09:07 crc kubenswrapper[4678]: E1013 13:09:07.593506 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:09:08 crc kubenswrapper[4678]: I1013 13:09:08.059486 4678 generic.go:334] "Generic (PLEG): container finished" podID="48c09433-04bb-4cd0-8bfd-72606ebb285a" containerID="d8289f6f629becc1f52dc2348102dc199c780a7399325fed4919921d91bfc113" exitCode=0 Oct 13 13:09:08 crc kubenswrapper[4678]: I1013 13:09:08.059553 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" event={"ID":"48c09433-04bb-4cd0-8bfd-72606ebb285a","Type":"ContainerDied","Data":"d8289f6f629becc1f52dc2348102dc199c780a7399325fed4919921d91bfc113"} Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.523179 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.620640 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7ts7\" (UniqueName: \"kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7\") pod \"48c09433-04bb-4cd0-8bfd-72606ebb285a\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.620788 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory\") pod \"48c09433-04bb-4cd0-8bfd-72606ebb285a\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.620839 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key\") pod \"48c09433-04bb-4cd0-8bfd-72606ebb285a\" (UID: \"48c09433-04bb-4cd0-8bfd-72606ebb285a\") " Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.630395 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7" (OuterVolumeSpecName: "kube-api-access-p7ts7") pod "48c09433-04bb-4cd0-8bfd-72606ebb285a" (UID: "48c09433-04bb-4cd0-8bfd-72606ebb285a"). InnerVolumeSpecName "kube-api-access-p7ts7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.645835 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory" (OuterVolumeSpecName: "inventory") pod "48c09433-04bb-4cd0-8bfd-72606ebb285a" (UID: "48c09433-04bb-4cd0-8bfd-72606ebb285a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.647194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48c09433-04bb-4cd0-8bfd-72606ebb285a" (UID: "48c09433-04bb-4cd0-8bfd-72606ebb285a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.722956 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.722982 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48c09433-04bb-4cd0-8bfd-72606ebb285a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:09:09 crc kubenswrapper[4678]: I1013 13:09:09.722991 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7ts7\" (UniqueName: \"kubernetes.io/projected/48c09433-04bb-4cd0-8bfd-72606ebb285a-kube-api-access-p7ts7\") on node \"crc\" DevicePath \"\"" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.078163 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" event={"ID":"48c09433-04bb-4cd0-8bfd-72606ebb285a","Type":"ContainerDied","Data":"da589256880105ea57a0f362ae9b47eb0068a5d0989b9cef75ee08e2f3632255"} Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.078206 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da589256880105ea57a0f362ae9b47eb0068a5d0989b9cef75ee08e2f3632255" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.078266 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hn894" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.170434 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z"] Oct 13 13:09:10 crc kubenswrapper[4678]: E1013 13:09:10.170828 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48c09433-04bb-4cd0-8bfd-72606ebb285a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.170852 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="48c09433-04bb-4cd0-8bfd-72606ebb285a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 13:09:10 crc kubenswrapper[4678]: E1013 13:09:10.170867 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="registry-server" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.170874 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="registry-server" Oct 13 13:09:10 crc kubenswrapper[4678]: E1013 13:09:10.170901 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="extract-utilities" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.170907 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="extract-utilities" Oct 13 13:09:10 crc kubenswrapper[4678]: E1013 13:09:10.170923 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="extract-content" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.170929 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="extract-content" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.171132 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="48c09433-04bb-4cd0-8bfd-72606ebb285a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.171145 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="68fb33a8-29dc-4468-9c4c-6b01e8d9a07e" containerName="registry-server" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.171799 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.174890 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.174919 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.176604 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.176816 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.180242 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z"] Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.231597 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.231668 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vgz\" (UniqueName: \"kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.232020 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.333859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.333956 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vgz\" (UniqueName: \"kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.334134 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.337667 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.338440 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.350451 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vgz\" (UniqueName: \"kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:10 crc kubenswrapper[4678]: I1013 13:09:10.492361 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:09:11 crc kubenswrapper[4678]: I1013 13:09:11.014606 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z"] Oct 13 13:09:11 crc kubenswrapper[4678]: W1013 13:09:11.015154 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0fd32ad_0b5d_4287_a5db_79c362e818e6.slice/crio-c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516 WatchSource:0}: Error finding container c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516: Status 404 returned error can't find the container with id c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516 Oct 13 13:09:11 crc kubenswrapper[4678]: I1013 13:09:11.017747 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:09:11 crc kubenswrapper[4678]: I1013 13:09:11.086895 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" event={"ID":"e0fd32ad-0b5d-4287-a5db-79c362e818e6","Type":"ContainerStarted","Data":"c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516"} Oct 13 13:09:12 crc kubenswrapper[4678]: I1013 13:09:12.032286 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-j9fqj"] Oct 13 13:09:12 crc kubenswrapper[4678]: I1013 13:09:12.042645 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-j9fqj"] Oct 13 13:09:12 crc kubenswrapper[4678]: I1013 13:09:12.095832 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" event={"ID":"e0fd32ad-0b5d-4287-a5db-79c362e818e6","Type":"ContainerStarted","Data":"e5dfc5cfa1b58701c18be300f3d2ba802c412b435e44eb4b90d348ec8978502b"} Oct 13 13:09:12 crc kubenswrapper[4678]: I1013 13:09:12.119280 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" podStartSLOduration=1.404255795 podStartE2EDuration="2.119260272s" podCreationTimestamp="2025-10-13 13:09:10 +0000 UTC" firstStartedPulling="2025-10-13 13:09:11.017568079 +0000 UTC m=+1539.102105963" lastFinishedPulling="2025-10-13 13:09:11.732572556 +0000 UTC m=+1539.817110440" observedRunningTime="2025-10-13 13:09:12.10799319 +0000 UTC m=+1540.192531064" watchObservedRunningTime="2025-10-13 13:09:12.119260272 +0000 UTC m=+1540.203798156" Oct 13 13:09:12 crc kubenswrapper[4678]: I1013 13:09:12.608792 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c40f306-6673-4695-86ed-32704f67754c" path="/var/lib/kubelet/pods/4c40f306-6673-4695-86ed-32704f67754c/volumes" Oct 13 13:09:21 crc kubenswrapper[4678]: I1013 13:09:21.593485 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:09:21 crc kubenswrapper[4678]: E1013 13:09:21.594296 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:09:22 crc kubenswrapper[4678]: I1013 13:09:22.033140 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jmfrc"] Oct 13 13:09:22 crc kubenswrapper[4678]: I1013 13:09:22.039258 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jmfrc"] Oct 13 13:09:22 crc kubenswrapper[4678]: I1013 13:09:22.607699 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3931313a-c89a-442e-b955-56aeeaf0d200" path="/var/lib/kubelet/pods/3931313a-c89a-442e-b955-56aeeaf0d200/volumes" Oct 13 13:09:33 crc kubenswrapper[4678]: I1013 13:09:33.592483 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:09:33 crc kubenswrapper[4678]: E1013 13:09:33.593252 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:09:45 crc kubenswrapper[4678]: I1013 13:09:45.592838 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:09:45 crc kubenswrapper[4678]: E1013 13:09:45.594090 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:09:51 crc kubenswrapper[4678]: I1013 13:09:51.061136 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-55g4b"] Oct 13 13:09:51 crc kubenswrapper[4678]: I1013 13:09:51.073765 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-55g4b"] Oct 13 13:09:52 crc kubenswrapper[4678]: I1013 13:09:52.608547 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b72b73c0-6cc4-48fe-a32f-784786617f79" path="/var/lib/kubelet/pods/b72b73c0-6cc4-48fe-a32f-784786617f79/volumes" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.729144 4678 scope.go:117] "RemoveContainer" containerID="a21d4cfbd1158a33b8fcbd4c06b615d24c72654d420145c0013dd93e67aed367" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.771714 4678 scope.go:117] "RemoveContainer" containerID="9a5fdeed6f438cf5638f1d8b942d9e37b0926110d365a83f936438e2cf76aa36" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.832659 4678 scope.go:117] "RemoveContainer" containerID="d41df3736e00c5b005c542f164030615478d676dcb300a5311cedea908a8c051" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.884674 4678 scope.go:117] "RemoveContainer" containerID="f64fa91748745d96b80fa05e4d45f8b497f118f1943296f5bd98b13a9929cd4a" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.926170 4678 scope.go:117] "RemoveContainer" containerID="6635145ecad9eb42dc99dd83ebfd65d3a214560efe747c3a5f49ea8354df55a3" Oct 13 13:09:53 crc kubenswrapper[4678]: I1013 13:09:53.988700 4678 scope.go:117] "RemoveContainer" containerID="23f30c99c952ff26c8254f54695d722b297d62ee3cc21813ba5a2e987e238eb4" Oct 13 13:09:56 crc kubenswrapper[4678]: I1013 13:09:56.593127 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:09:56 crc kubenswrapper[4678]: E1013 13:09:56.593841 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:09:58 crc kubenswrapper[4678]: I1013 13:09:58.038010 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jb889"] Oct 13 13:09:58 crc kubenswrapper[4678]: I1013 13:09:58.047816 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jb889"] Oct 13 13:09:58 crc kubenswrapper[4678]: I1013 13:09:58.604714 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b7cde62-fcc9-40d6-8099-58838aea1ea7" path="/var/lib/kubelet/pods/5b7cde62-fcc9-40d6-8099-58838aea1ea7/volumes" Oct 13 13:10:01 crc kubenswrapper[4678]: I1013 13:10:01.030320 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xldvz"] Oct 13 13:10:01 crc kubenswrapper[4678]: I1013 13:10:01.045469 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xldvz"] Oct 13 13:10:02 crc kubenswrapper[4678]: I1013 13:10:02.615712 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c27b4c-8e65-490b-8ef7-aba250924b92" path="/var/lib/kubelet/pods/20c27b4c-8e65-490b-8ef7-aba250924b92/volumes" Oct 13 13:10:08 crc kubenswrapper[4678]: I1013 13:10:08.592420 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:10:08 crc kubenswrapper[4678]: E1013 13:10:08.593686 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:10:10 crc kubenswrapper[4678]: I1013 13:10:10.029226 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-p9mq7"] Oct 13 13:10:10 crc kubenswrapper[4678]: I1013 13:10:10.038914 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-p9mq7"] Oct 13 13:10:10 crc kubenswrapper[4678]: I1013 13:10:10.604521 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e988f0-2dc8-47d9-a8e0-65550e610593" path="/var/lib/kubelet/pods/04e988f0-2dc8-47d9-a8e0-65550e610593/volumes" Oct 13 13:10:17 crc kubenswrapper[4678]: I1013 13:10:17.030275 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-l6mcn"] Oct 13 13:10:17 crc kubenswrapper[4678]: I1013 13:10:17.038709 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-l6mcn"] Oct 13 13:10:18 crc kubenswrapper[4678]: I1013 13:10:18.607112 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="909be1eb-5445-4411-a2d0-c9e2618dd6bc" path="/var/lib/kubelet/pods/909be1eb-5445-4411-a2d0-c9e2618dd6bc/volumes" Oct 13 13:10:22 crc kubenswrapper[4678]: I1013 13:10:22.598202 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:10:22 crc kubenswrapper[4678]: E1013 13:10:22.599083 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:10:24 crc kubenswrapper[4678]: I1013 13:10:24.852802 4678 generic.go:334] "Generic (PLEG): container finished" podID="e0fd32ad-0b5d-4287-a5db-79c362e818e6" containerID="e5dfc5cfa1b58701c18be300f3d2ba802c412b435e44eb4b90d348ec8978502b" exitCode=0 Oct 13 13:10:24 crc kubenswrapper[4678]: I1013 13:10:24.852959 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" event={"ID":"e0fd32ad-0b5d-4287-a5db-79c362e818e6","Type":"ContainerDied","Data":"e5dfc5cfa1b58701c18be300f3d2ba802c412b435e44eb4b90d348ec8978502b"} Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.322424 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.349477 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key\") pod \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.349583 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vgz\" (UniqueName: \"kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz\") pod \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.349671 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory\") pod \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\" (UID: \"e0fd32ad-0b5d-4287-a5db-79c362e818e6\") " Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.357468 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz" (OuterVolumeSpecName: "kube-api-access-m7vgz") pod "e0fd32ad-0b5d-4287-a5db-79c362e818e6" (UID: "e0fd32ad-0b5d-4287-a5db-79c362e818e6"). InnerVolumeSpecName "kube-api-access-m7vgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.382995 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory" (OuterVolumeSpecName: "inventory") pod "e0fd32ad-0b5d-4287-a5db-79c362e818e6" (UID: "e0fd32ad-0b5d-4287-a5db-79c362e818e6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.401442 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e0fd32ad-0b5d-4287-a5db-79c362e818e6" (UID: "e0fd32ad-0b5d-4287-a5db-79c362e818e6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.452407 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.452458 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vgz\" (UniqueName: \"kubernetes.io/projected/e0fd32ad-0b5d-4287-a5db-79c362e818e6-kube-api-access-m7vgz\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.452476 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0fd32ad-0b5d-4287-a5db-79c362e818e6-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.877447 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" event={"ID":"e0fd32ad-0b5d-4287-a5db-79c362e818e6","Type":"ContainerDied","Data":"c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516"} Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.878127 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5e1ae3e0be4264b1219bfa0551ead0ce78f844c5a8550ae2279ef04c8a24516" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.877568 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.978601 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln"] Oct 13 13:10:26 crc kubenswrapper[4678]: E1013 13:10:26.979085 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0fd32ad-0b5d-4287-a5db-79c362e818e6" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.979103 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0fd32ad-0b5d-4287-a5db-79c362e818e6" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.979325 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0fd32ad-0b5d-4287-a5db-79c362e818e6" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.980071 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.984360 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.984592 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.984755 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.984898 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:10:26 crc kubenswrapper[4678]: I1013 13:10:26.996792 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln"] Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.067939 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.068270 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.068340 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4brm\" (UniqueName: \"kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.171285 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.171475 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.171512 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4brm\" (UniqueName: \"kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.176456 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.177223 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.189941 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4brm\" (UniqueName: \"kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sqqln\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.345624 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:27 crc kubenswrapper[4678]: I1013 13:10:27.921187 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln"] Oct 13 13:10:28 crc kubenswrapper[4678]: I1013 13:10:28.903258 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" event={"ID":"5fc607e5-2f31-42b9-ba24-ba02b1516a75","Type":"ContainerStarted","Data":"303d01d264d863e6f4eb2fe89f8e77cf821c64506f770d9a460eafbefa7e812d"} Oct 13 13:10:28 crc kubenswrapper[4678]: I1013 13:10:28.903976 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" event={"ID":"5fc607e5-2f31-42b9-ba24-ba02b1516a75","Type":"ContainerStarted","Data":"9c144789e8ceb39f940132a8393050bd85f8f6d95e2e76570c2ab06219e71ecf"} Oct 13 13:10:28 crc kubenswrapper[4678]: I1013 13:10:28.924546 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" podStartSLOduration=2.300611334 podStartE2EDuration="2.924527046s" podCreationTimestamp="2025-10-13 13:10:26 +0000 UTC" firstStartedPulling="2025-10-13 13:10:27.929370969 +0000 UTC m=+1616.013908853" lastFinishedPulling="2025-10-13 13:10:28.553286651 +0000 UTC m=+1616.637824565" observedRunningTime="2025-10-13 13:10:28.922917815 +0000 UTC m=+1617.007455699" watchObservedRunningTime="2025-10-13 13:10:28.924527046 +0000 UTC m=+1617.009064950" Oct 13 13:10:33 crc kubenswrapper[4678]: I1013 13:10:33.946304 4678 generic.go:334] "Generic (PLEG): container finished" podID="5fc607e5-2f31-42b9-ba24-ba02b1516a75" containerID="303d01d264d863e6f4eb2fe89f8e77cf821c64506f770d9a460eafbefa7e812d" exitCode=0 Oct 13 13:10:33 crc kubenswrapper[4678]: I1013 13:10:33.946366 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" event={"ID":"5fc607e5-2f31-42b9-ba24-ba02b1516a75","Type":"ContainerDied","Data":"303d01d264d863e6f4eb2fe89f8e77cf821c64506f770d9a460eafbefa7e812d"} Oct 13 13:10:34 crc kubenswrapper[4678]: I1013 13:10:34.592706 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:10:34 crc kubenswrapper[4678]: E1013 13:10:34.593476 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.392474 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.453836 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4brm\" (UniqueName: \"kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm\") pod \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.453963 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key\") pod \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.454028 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory\") pod \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\" (UID: \"5fc607e5-2f31-42b9-ba24-ba02b1516a75\") " Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.459727 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm" (OuterVolumeSpecName: "kube-api-access-g4brm") pod "5fc607e5-2f31-42b9-ba24-ba02b1516a75" (UID: "5fc607e5-2f31-42b9-ba24-ba02b1516a75"). InnerVolumeSpecName "kube-api-access-g4brm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.481439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory" (OuterVolumeSpecName: "inventory") pod "5fc607e5-2f31-42b9-ba24-ba02b1516a75" (UID: "5fc607e5-2f31-42b9-ba24-ba02b1516a75"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.483681 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fc607e5-2f31-42b9-ba24-ba02b1516a75" (UID: "5fc607e5-2f31-42b9-ba24-ba02b1516a75"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.556786 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4brm\" (UniqueName: \"kubernetes.io/projected/5fc607e5-2f31-42b9-ba24-ba02b1516a75-kube-api-access-g4brm\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.556826 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.556837 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fc607e5-2f31-42b9-ba24-ba02b1516a75-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.968559 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" event={"ID":"5fc607e5-2f31-42b9-ba24-ba02b1516a75","Type":"ContainerDied","Data":"9c144789e8ceb39f940132a8393050bd85f8f6d95e2e76570c2ab06219e71ecf"} Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.968597 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c144789e8ceb39f940132a8393050bd85f8f6d95e2e76570c2ab06219e71ecf" Oct 13 13:10:35 crc kubenswrapper[4678]: I1013 13:10:35.968773 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sqqln" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.136719 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt"] Oct 13 13:10:36 crc kubenswrapper[4678]: E1013 13:10:36.137169 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc607e5-2f31-42b9-ba24-ba02b1516a75" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.137191 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc607e5-2f31-42b9-ba24-ba02b1516a75" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.137442 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc607e5-2f31-42b9-ba24-ba02b1516a75" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.138192 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.144701 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.144739 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.145259 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.145381 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.150290 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt"] Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.170590 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2v8n\" (UniqueName: \"kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.170675 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.170703 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.272152 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2v8n\" (UniqueName: \"kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.272506 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.272533 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.276296 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.281533 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.294170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2v8n\" (UniqueName: \"kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-ssndt\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:36 crc kubenswrapper[4678]: I1013 13:10:36.462044 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:10:37 crc kubenswrapper[4678]: I1013 13:10:37.016829 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt"] Oct 13 13:10:37 crc kubenswrapper[4678]: I1013 13:10:37.989163 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" event={"ID":"84dfdaf8-938e-4eca-af31-79b85a483599","Type":"ContainerStarted","Data":"447c20c13097da17e8d248f02d6125a67ad731f3c3dbb7fa6cbcc21eace4cfe2"} Oct 13 13:10:37 crc kubenswrapper[4678]: I1013 13:10:37.989549 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" event={"ID":"84dfdaf8-938e-4eca-af31-79b85a483599","Type":"ContainerStarted","Data":"cfc1b02e9ab4842cf2de061d9e1e990d1c5d388e951511e26f3855d846d0d984"} Oct 13 13:10:38 crc kubenswrapper[4678]: I1013 13:10:38.019174 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" podStartSLOduration=1.596159618 podStartE2EDuration="2.019150906s" podCreationTimestamp="2025-10-13 13:10:36 +0000 UTC" firstStartedPulling="2025-10-13 13:10:37.015618501 +0000 UTC m=+1625.100156385" lastFinishedPulling="2025-10-13 13:10:37.438609779 +0000 UTC m=+1625.523147673" observedRunningTime="2025-10-13 13:10:38.002463583 +0000 UTC m=+1626.087001477" watchObservedRunningTime="2025-10-13 13:10:38.019150906 +0000 UTC m=+1626.103688820" Oct 13 13:10:45 crc kubenswrapper[4678]: I1013 13:10:45.592971 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:10:45 crc kubenswrapper[4678]: E1013 13:10:45.594600 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.040350 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ftqv2"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.051893 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xmd6j"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.063987 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-nhm8n"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.073075 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ftqv2"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.079969 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xmd6j"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.087580 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-nhm8n"] Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.608655 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aac9187-4105-4d36-b4f0-ad03156a500f" path="/var/lib/kubelet/pods/4aac9187-4105-4d36-b4f0-ad03156a500f/volumes" Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.610242 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc51676-abd0-46b9-8bde-89b0f8d501ce" path="/var/lib/kubelet/pods/4bc51676-abd0-46b9-8bde-89b0f8d501ce/volumes" Oct 13 13:10:52 crc kubenswrapper[4678]: I1013 13:10:52.612221 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bbe76ec-9cc8-4bf1-bda6-ccf07a521616" path="/var/lib/kubelet/pods/6bbe76ec-9cc8-4bf1-bda6-ccf07a521616/volumes" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.166785 4678 scope.go:117] "RemoveContainer" containerID="5e60550dd91d7361c7c27c077becba979569b6bca27c6b1e3e2b14b7ee09b894" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.199603 4678 scope.go:117] "RemoveContainer" containerID="bafc36f737e2055718d77357bfccba9a4187b0378c6d7b66cdd3a9b8fdcec96c" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.289686 4678 scope.go:117] "RemoveContainer" containerID="17cc3b99ef9dc7efbe47e41d176a3fc64b8839ad186c3dbb080c967953deecc0" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.316685 4678 scope.go:117] "RemoveContainer" containerID="dc3b37825d146cb808e7264a0b62e84dc6031e46bd732274f224c450da6148aa" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.393199 4678 scope.go:117] "RemoveContainer" containerID="b361fadc7b71f790f9e8923b00536175892297f49ef4481560a3a3fb3d55374b" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.431592 4678 scope.go:117] "RemoveContainer" containerID="9834e963784cdd5abdcd409fcc1d1623f73c73ed6a621eb5f991c2f5ec833da6" Oct 13 13:10:54 crc kubenswrapper[4678]: I1013 13:10:54.464645 4678 scope.go:117] "RemoveContainer" containerID="083dc5b008aea5e16e3698c661030e91b999e9621bb64def235fa07d4fd8ed8b" Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.036415 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8b33-account-create-597qt"] Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.047312 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-6afa-account-create-sf9g6"] Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.056899 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8b33-account-create-597qt"] Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.073484 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-6afa-account-create-sf9g6"] Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.592189 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:10:58 crc kubenswrapper[4678]: E1013 13:10:58.592466 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.601803 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16d1d3ee-bad0-4228-801e-9f8e14c9170c" path="/var/lib/kubelet/pods/16d1d3ee-bad0-4228-801e-9f8e14c9170c/volumes" Oct 13 13:10:58 crc kubenswrapper[4678]: I1013 13:10:58.602518 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4d469c9-e858-4fe1-a158-2828b9c35667" path="/var/lib/kubelet/pods/b4d469c9-e858-4fe1-a158-2828b9c35667/volumes" Oct 13 13:11:11 crc kubenswrapper[4678]: I1013 13:11:11.028611 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-df52-account-create-qdxzz"] Oct 13 13:11:11 crc kubenswrapper[4678]: I1013 13:11:11.035264 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-df52-account-create-qdxzz"] Oct 13 13:11:12 crc kubenswrapper[4678]: I1013 13:11:12.605919 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ba84a93-4108-4b9b-85c0-e4e0e8219398" path="/var/lib/kubelet/pods/5ba84a93-4108-4b9b-85c0-e4e0e8219398/volumes" Oct 13 13:11:13 crc kubenswrapper[4678]: I1013 13:11:13.593890 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:11:13 crc kubenswrapper[4678]: E1013 13:11:13.594181 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:11:14 crc kubenswrapper[4678]: I1013 13:11:14.355999 4678 generic.go:334] "Generic (PLEG): container finished" podID="84dfdaf8-938e-4eca-af31-79b85a483599" containerID="447c20c13097da17e8d248f02d6125a67ad731f3c3dbb7fa6cbcc21eace4cfe2" exitCode=0 Oct 13 13:11:14 crc kubenswrapper[4678]: I1013 13:11:14.356115 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" event={"ID":"84dfdaf8-938e-4eca-af31-79b85a483599","Type":"ContainerDied","Data":"447c20c13097da17e8d248f02d6125a67ad731f3c3dbb7fa6cbcc21eace4cfe2"} Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.758077 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.798009 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key\") pod \"84dfdaf8-938e-4eca-af31-79b85a483599\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.798180 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2v8n\" (UniqueName: \"kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n\") pod \"84dfdaf8-938e-4eca-af31-79b85a483599\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.798240 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory\") pod \"84dfdaf8-938e-4eca-af31-79b85a483599\" (UID: \"84dfdaf8-938e-4eca-af31-79b85a483599\") " Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.812688 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n" (OuterVolumeSpecName: "kube-api-access-s2v8n") pod "84dfdaf8-938e-4eca-af31-79b85a483599" (UID: "84dfdaf8-938e-4eca-af31-79b85a483599"). InnerVolumeSpecName "kube-api-access-s2v8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.830744 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "84dfdaf8-938e-4eca-af31-79b85a483599" (UID: "84dfdaf8-938e-4eca-af31-79b85a483599"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.834264 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory" (OuterVolumeSpecName: "inventory") pod "84dfdaf8-938e-4eca-af31-79b85a483599" (UID: "84dfdaf8-938e-4eca-af31-79b85a483599"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.900366 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.900401 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2v8n\" (UniqueName: \"kubernetes.io/projected/84dfdaf8-938e-4eca-af31-79b85a483599-kube-api-access-s2v8n\") on node \"crc\" DevicePath \"\"" Oct 13 13:11:15 crc kubenswrapper[4678]: I1013 13:11:15.900413 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/84dfdaf8-938e-4eca-af31-79b85a483599-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.374028 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" event={"ID":"84dfdaf8-938e-4eca-af31-79b85a483599","Type":"ContainerDied","Data":"cfc1b02e9ab4842cf2de061d9e1e990d1c5d388e951511e26f3855d846d0d984"} Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.374411 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfc1b02e9ab4842cf2de061d9e1e990d1c5d388e951511e26f3855d846d0d984" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.374133 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-ssndt" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.456873 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz"] Oct 13 13:11:16 crc kubenswrapper[4678]: E1013 13:11:16.457641 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84dfdaf8-938e-4eca-af31-79b85a483599" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.457672 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="84dfdaf8-938e-4eca-af31-79b85a483599" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.457938 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="84dfdaf8-938e-4eca-af31-79b85a483599" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.458757 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.461657 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.462010 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.462013 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.462713 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.467152 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz"] Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.510635 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.510865 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.511016 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zn4b\" (UniqueName: \"kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.612937 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.613173 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.613221 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zn4b\" (UniqueName: \"kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.618161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.619306 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.635555 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zn4b\" (UniqueName: \"kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:16 crc kubenswrapper[4678]: I1013 13:11:16.788131 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:11:17 crc kubenswrapper[4678]: I1013 13:11:17.317143 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz"] Oct 13 13:11:17 crc kubenswrapper[4678]: I1013 13:11:17.382315 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" event={"ID":"567d50a7-a8dd-4bdb-82dc-8bee34a93d25","Type":"ContainerStarted","Data":"251c87745a6f9c972633cb5c27643ba35400bc8c2f36e6a4cad97ad71542e80b"} Oct 13 13:11:18 crc kubenswrapper[4678]: I1013 13:11:18.394354 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" event={"ID":"567d50a7-a8dd-4bdb-82dc-8bee34a93d25","Type":"ContainerStarted","Data":"25c09957ed95e607da2a97bf3790e1c431f711ac677bb622b903031c61e6fa07"} Oct 13 13:11:18 crc kubenswrapper[4678]: I1013 13:11:18.418348 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" podStartSLOduration=1.8781305640000001 podStartE2EDuration="2.418333164s" podCreationTimestamp="2025-10-13 13:11:16 +0000 UTC" firstStartedPulling="2025-10-13 13:11:17.319808174 +0000 UTC m=+1665.404346058" lastFinishedPulling="2025-10-13 13:11:17.860010754 +0000 UTC m=+1665.944548658" observedRunningTime="2025-10-13 13:11:18.408106619 +0000 UTC m=+1666.492644503" watchObservedRunningTime="2025-10-13 13:11:18.418333164 +0000 UTC m=+1666.502871048" Oct 13 13:11:23 crc kubenswrapper[4678]: I1013 13:11:23.044969 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7l6s4"] Oct 13 13:11:23 crc kubenswrapper[4678]: I1013 13:11:23.052815 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7l6s4"] Oct 13 13:11:24 crc kubenswrapper[4678]: I1013 13:11:24.593548 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:11:24 crc kubenswrapper[4678]: E1013 13:11:24.594779 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:11:24 crc kubenswrapper[4678]: I1013 13:11:24.602326 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b413c56-81e8-4b62-aa08-742ceca80ec7" path="/var/lib/kubelet/pods/2b413c56-81e8-4b62-aa08-742ceca80ec7/volumes" Oct 13 13:11:38 crc kubenswrapper[4678]: I1013 13:11:38.593799 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:11:38 crc kubenswrapper[4678]: E1013 13:11:38.594568 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:11:44 crc kubenswrapper[4678]: I1013 13:11:44.037922 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-m7csp"] Oct 13 13:11:44 crc kubenswrapper[4678]: I1013 13:11:44.052595 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-m7csp"] Oct 13 13:11:44 crc kubenswrapper[4678]: I1013 13:11:44.602876 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa05290-ebc1-4886-b08c-0ac9b8c314f0" path="/var/lib/kubelet/pods/9aa05290-ebc1-4886-b08c-0ac9b8c314f0/volumes" Oct 13 13:11:45 crc kubenswrapper[4678]: I1013 13:11:45.038374 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dpn5h"] Oct 13 13:11:45 crc kubenswrapper[4678]: I1013 13:11:45.051883 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dpn5h"] Oct 13 13:11:46 crc kubenswrapper[4678]: I1013 13:11:46.604529 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b705be0b-2f5e-4778-8160-4f9863ccbda1" path="/var/lib/kubelet/pods/b705be0b-2f5e-4778-8160-4f9863ccbda1/volumes" Oct 13 13:11:50 crc kubenswrapper[4678]: I1013 13:11:50.592828 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:11:50 crc kubenswrapper[4678]: E1013 13:11:50.593622 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:11:54 crc kubenswrapper[4678]: I1013 13:11:54.629888 4678 scope.go:117] "RemoveContainer" containerID="c8f9f61de16d10814568f1738a0516ee253d9c78f8ccf604d475171b73fe995b" Oct 13 13:11:54 crc kubenswrapper[4678]: I1013 13:11:54.678861 4678 scope.go:117] "RemoveContainer" containerID="249bac6a4adbf8689a245ee9be7da465ae58728ba47e9dccf3f685be9e1ea5d4" Oct 13 13:11:54 crc kubenswrapper[4678]: I1013 13:11:54.757588 4678 scope.go:117] "RemoveContainer" containerID="aebb155e8ca19274c5d9798de4133f40c054a7e10fff3363cdf162e4be5149af" Oct 13 13:11:54 crc kubenswrapper[4678]: I1013 13:11:54.827736 4678 scope.go:117] "RemoveContainer" containerID="71ffbb6871d897e5e68bd12eb6a0988f656723dafd788f5cadb5d2e21068a734" Oct 13 13:11:54 crc kubenswrapper[4678]: I1013 13:11:54.878114 4678 scope.go:117] "RemoveContainer" containerID="e9bdcaa3ec8c11fb225963bb6505e33e00b65e603ae23c015784a9ce8aea1491" Oct 13 13:11:55 crc kubenswrapper[4678]: I1013 13:11:55.023579 4678 scope.go:117] "RemoveContainer" containerID="6993e5d1e6466cefdfa6ff4770a9fe3bce7b274dc3659f6402018d178abe7121" Oct 13 13:12:02 crc kubenswrapper[4678]: I1013 13:12:02.603589 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:12:02 crc kubenswrapper[4678]: E1013 13:12:02.604786 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:12:12 crc kubenswrapper[4678]: I1013 13:12:12.887773 4678 generic.go:334] "Generic (PLEG): container finished" podID="567d50a7-a8dd-4bdb-82dc-8bee34a93d25" containerID="25c09957ed95e607da2a97bf3790e1c431f711ac677bb622b903031c61e6fa07" exitCode=2 Oct 13 13:12:12 crc kubenswrapper[4678]: I1013 13:12:12.887846 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" event={"ID":"567d50a7-a8dd-4bdb-82dc-8bee34a93d25","Type":"ContainerDied","Data":"25c09957ed95e607da2a97bf3790e1c431f711ac677bb622b903031c61e6fa07"} Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.371792 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.536904 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory\") pod \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.537108 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zn4b\" (UniqueName: \"kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b\") pod \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.537325 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key\") pod \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\" (UID: \"567d50a7-a8dd-4bdb-82dc-8bee34a93d25\") " Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.544263 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b" (OuterVolumeSpecName: "kube-api-access-6zn4b") pod "567d50a7-a8dd-4bdb-82dc-8bee34a93d25" (UID: "567d50a7-a8dd-4bdb-82dc-8bee34a93d25"). InnerVolumeSpecName "kube-api-access-6zn4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.577210 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "567d50a7-a8dd-4bdb-82dc-8bee34a93d25" (UID: "567d50a7-a8dd-4bdb-82dc-8bee34a93d25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.585199 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory" (OuterVolumeSpecName: "inventory") pod "567d50a7-a8dd-4bdb-82dc-8bee34a93d25" (UID: "567d50a7-a8dd-4bdb-82dc-8bee34a93d25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.639679 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.639718 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.639730 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zn4b\" (UniqueName: \"kubernetes.io/projected/567d50a7-a8dd-4bdb-82dc-8bee34a93d25-kube-api-access-6zn4b\") on node \"crc\" DevicePath \"\"" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.910501 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" event={"ID":"567d50a7-a8dd-4bdb-82dc-8bee34a93d25","Type":"ContainerDied","Data":"251c87745a6f9c972633cb5c27643ba35400bc8c2f36e6a4cad97ad71542e80b"} Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.910886 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="251c87745a6f9c972633cb5c27643ba35400bc8c2f36e6a4cad97ad71542e80b" Oct 13 13:12:14 crc kubenswrapper[4678]: I1013 13:12:14.910587 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz" Oct 13 13:12:17 crc kubenswrapper[4678]: I1013 13:12:17.594175 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:12:17 crc kubenswrapper[4678]: E1013 13:12:17.595085 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.026014 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd"] Oct 13 13:12:22 crc kubenswrapper[4678]: E1013 13:12:22.026613 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="567d50a7-a8dd-4bdb-82dc-8bee34a93d25" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.026625 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="567d50a7-a8dd-4bdb-82dc-8bee34a93d25" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.026836 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="567d50a7-a8dd-4bdb-82dc-8bee34a93d25" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.027591 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.029708 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.031283 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.031498 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.031591 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.037099 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd"] Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.079150 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf74w\" (UniqueName: \"kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.079225 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.079265 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.182531 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf74w\" (UniqueName: \"kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.182636 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.182689 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.191801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.194289 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.201834 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf74w\" (UniqueName: \"kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.355850 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.907625 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd"] Oct 13 13:12:22 crc kubenswrapper[4678]: I1013 13:12:22.986229 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" event={"ID":"0dacde92-4685-4950-8219-d634cbd0b1cc","Type":"ContainerStarted","Data":"d9767644b1998ef3fb6ac6d62318f91b7f1999dabf3103ad015d8cb4188c6e98"} Oct 13 13:12:24 crc kubenswrapper[4678]: I1013 13:12:24.000702 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" event={"ID":"0dacde92-4685-4950-8219-d634cbd0b1cc","Type":"ContainerStarted","Data":"8af5bd8325f1df50d5057669bdcbe00b8f50317ee3da0de6e5d2000ed09cd1c5"} Oct 13 13:12:24 crc kubenswrapper[4678]: I1013 13:12:24.029066 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" podStartSLOduration=1.456922444 podStartE2EDuration="2.02903042s" podCreationTimestamp="2025-10-13 13:12:22 +0000 UTC" firstStartedPulling="2025-10-13 13:12:22.910932836 +0000 UTC m=+1730.995470720" lastFinishedPulling="2025-10-13 13:12:23.483040812 +0000 UTC m=+1731.567578696" observedRunningTime="2025-10-13 13:12:24.02131624 +0000 UTC m=+1732.105854204" watchObservedRunningTime="2025-10-13 13:12:24.02903042 +0000 UTC m=+1732.113568324" Oct 13 13:12:29 crc kubenswrapper[4678]: I1013 13:12:29.037018 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jjdj"] Oct 13 13:12:29 crc kubenswrapper[4678]: I1013 13:12:29.047188 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jjdj"] Oct 13 13:12:29 crc kubenswrapper[4678]: I1013 13:12:29.594678 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:12:29 crc kubenswrapper[4678]: E1013 13:12:29.595334 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:12:30 crc kubenswrapper[4678]: I1013 13:12:30.606770 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2afc7094-0e68-469c-877a-5ce77ce47807" path="/var/lib/kubelet/pods/2afc7094-0e68-469c-877a-5ce77ce47807/volumes" Oct 13 13:12:43 crc kubenswrapper[4678]: I1013 13:12:43.593817 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:12:43 crc kubenswrapper[4678]: E1013 13:12:43.594625 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:12:54 crc kubenswrapper[4678]: I1013 13:12:54.592771 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:12:54 crc kubenswrapper[4678]: E1013 13:12:54.593943 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.142804 4678 scope.go:117] "RemoveContainer" containerID="4f7c25f8c1f5e4ca130fbef0a63a51f0c96899dc07dd9141356fada68d354bf7" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.841488 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.843271 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.859618 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.879488 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.879608 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtxtt\" (UniqueName: \"kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.882405 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.985083 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.985192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtxtt\" (UniqueName: \"kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.985398 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.985782 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:55 crc kubenswrapper[4678]: I1013 13:12:55.985894 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:56 crc kubenswrapper[4678]: I1013 13:12:56.008747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtxtt\" (UniqueName: \"kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt\") pod \"redhat-marketplace-fd422\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:56 crc kubenswrapper[4678]: I1013 13:12:56.168205 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:12:56 crc kubenswrapper[4678]: I1013 13:12:56.636778 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:12:56 crc kubenswrapper[4678]: W1013 13:12:56.647316 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9733c8b_d7a8_4dd6_b6f4_1443ad2c0601.slice/crio-f843f386c1c1bc012fd416c7f7570a1c2a0317328d960d472c4c897510d91e05 WatchSource:0}: Error finding container f843f386c1c1bc012fd416c7f7570a1c2a0317328d960d472c4c897510d91e05: Status 404 returned error can't find the container with id f843f386c1c1bc012fd416c7f7570a1c2a0317328d960d472c4c897510d91e05 Oct 13 13:12:57 crc kubenswrapper[4678]: I1013 13:12:57.347371 4678 generic.go:334] "Generic (PLEG): container finished" podID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerID="688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353" exitCode=0 Oct 13 13:12:57 crc kubenswrapper[4678]: I1013 13:12:57.347760 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerDied","Data":"688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353"} Oct 13 13:12:57 crc kubenswrapper[4678]: I1013 13:12:57.347794 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerStarted","Data":"f843f386c1c1bc012fd416c7f7570a1c2a0317328d960d472c4c897510d91e05"} Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.356214 4678 generic.go:334] "Generic (PLEG): container finished" podID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerID="245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724" exitCode=0 Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.356337 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerDied","Data":"245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724"} Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.439829 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.442786 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.451749 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.636631 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.636693 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9ddg\" (UniqueName: \"kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.636773 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.739080 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.739227 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.739273 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9ddg\" (UniqueName: \"kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.739547 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.739828 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.760249 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9ddg\" (UniqueName: \"kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg\") pod \"redhat-operators-844xn\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:58 crc kubenswrapper[4678]: I1013 13:12:58.763595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.059850 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.366204 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerID="641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1" exitCode=0 Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.366355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerDied","Data":"641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1"} Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.366548 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerStarted","Data":"11b2970ac500f8fb89d8757c8aa870041f27882b0ed4633f5f84ef38734f6fd2"} Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.368942 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerStarted","Data":"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6"} Oct 13 13:12:59 crc kubenswrapper[4678]: I1013 13:12:59.408315 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fd422" podStartSLOduration=2.787399842 podStartE2EDuration="4.408297721s" podCreationTimestamp="2025-10-13 13:12:55 +0000 UTC" firstStartedPulling="2025-10-13 13:12:57.351035839 +0000 UTC m=+1765.435573723" lastFinishedPulling="2025-10-13 13:12:58.971933718 +0000 UTC m=+1767.056471602" observedRunningTime="2025-10-13 13:12:59.401223188 +0000 UTC m=+1767.485761072" watchObservedRunningTime="2025-10-13 13:12:59.408297721 +0000 UTC m=+1767.492835605" Oct 13 13:13:01 crc kubenswrapper[4678]: I1013 13:13:01.389025 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerStarted","Data":"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb"} Oct 13 13:13:02 crc kubenswrapper[4678]: I1013 13:13:02.402666 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerID="5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb" exitCode=0 Oct 13 13:13:02 crc kubenswrapper[4678]: I1013 13:13:02.402721 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerDied","Data":"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb"} Oct 13 13:13:04 crc kubenswrapper[4678]: I1013 13:13:04.423704 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerStarted","Data":"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca"} Oct 13 13:13:04 crc kubenswrapper[4678]: I1013 13:13:04.441977 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-844xn" podStartSLOduration=2.413344131 podStartE2EDuration="6.441955747s" podCreationTimestamp="2025-10-13 13:12:58 +0000 UTC" firstStartedPulling="2025-10-13 13:12:59.368087018 +0000 UTC m=+1767.452624902" lastFinishedPulling="2025-10-13 13:13:03.396698624 +0000 UTC m=+1771.481236518" observedRunningTime="2025-10-13 13:13:04.440186961 +0000 UTC m=+1772.524724845" watchObservedRunningTime="2025-10-13 13:13:04.441955747 +0000 UTC m=+1772.526493631" Oct 13 13:13:06 crc kubenswrapper[4678]: I1013 13:13:06.168882 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:06 crc kubenswrapper[4678]: I1013 13:13:06.169308 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:06 crc kubenswrapper[4678]: I1013 13:13:06.222929 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:06 crc kubenswrapper[4678]: I1013 13:13:06.495799 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:06 crc kubenswrapper[4678]: I1013 13:13:06.831230 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:13:07 crc kubenswrapper[4678]: I1013 13:13:07.458250 4678 generic.go:334] "Generic (PLEG): container finished" podID="0dacde92-4685-4950-8219-d634cbd0b1cc" containerID="8af5bd8325f1df50d5057669bdcbe00b8f50317ee3da0de6e5d2000ed09cd1c5" exitCode=0 Oct 13 13:13:07 crc kubenswrapper[4678]: I1013 13:13:07.458347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" event={"ID":"0dacde92-4685-4950-8219-d634cbd0b1cc","Type":"ContainerDied","Data":"8af5bd8325f1df50d5057669bdcbe00b8f50317ee3da0de6e5d2000ed09cd1c5"} Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.465904 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fd422" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="registry-server" containerID="cri-o://af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6" gracePeriod=2 Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.764642 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.765156 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.935103 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.941365 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945259 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content\") pod \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945303 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory\") pod \"0dacde92-4685-4950-8219-d634cbd0b1cc\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945332 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities\") pod \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945458 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf74w\" (UniqueName: \"kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w\") pod \"0dacde92-4685-4950-8219-d634cbd0b1cc\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945502 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key\") pod \"0dacde92-4685-4950-8219-d634cbd0b1cc\" (UID: \"0dacde92-4685-4950-8219-d634cbd0b1cc\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.945519 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtxtt\" (UniqueName: \"kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt\") pod \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\" (UID: \"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601\") " Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.948689 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities" (OuterVolumeSpecName: "utilities") pod "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" (UID: "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.953226 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt" (OuterVolumeSpecName: "kube-api-access-mtxtt") pod "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" (UID: "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601"). InnerVolumeSpecName "kube-api-access-mtxtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.953817 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w" (OuterVolumeSpecName: "kube-api-access-wf74w") pod "0dacde92-4685-4950-8219-d634cbd0b1cc" (UID: "0dacde92-4685-4950-8219-d634cbd0b1cc"). InnerVolumeSpecName "kube-api-access-wf74w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.964707 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" (UID: "e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.984984 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory" (OuterVolumeSpecName: "inventory") pod "0dacde92-4685-4950-8219-d634cbd0b1cc" (UID: "0dacde92-4685-4950-8219-d634cbd0b1cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:08 crc kubenswrapper[4678]: I1013 13:13:08.992881 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0dacde92-4685-4950-8219-d634cbd0b1cc" (UID: "0dacde92-4685-4950-8219-d634cbd0b1cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047326 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047359 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047368 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047377 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf74w\" (UniqueName: \"kubernetes.io/projected/0dacde92-4685-4950-8219-d634cbd0b1cc-kube-api-access-wf74w\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047387 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0dacde92-4685-4950-8219-d634cbd0b1cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.047397 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtxtt\" (UniqueName: \"kubernetes.io/projected/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601-kube-api-access-mtxtt\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.474981 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" event={"ID":"0dacde92-4685-4950-8219-d634cbd0b1cc","Type":"ContainerDied","Data":"d9767644b1998ef3fb6ac6d62318f91b7f1999dabf3103ad015d8cb4188c6e98"} Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.475373 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9767644b1998ef3fb6ac6d62318f91b7f1999dabf3103ad015d8cb4188c6e98" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.475002 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.479022 4678 generic.go:334] "Generic (PLEG): container finished" podID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerID="af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6" exitCode=0 Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.479113 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerDied","Data":"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6"} Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.479148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fd422" event={"ID":"e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601","Type":"ContainerDied","Data":"f843f386c1c1bc012fd416c7f7570a1c2a0317328d960d472c4c897510d91e05"} Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.479169 4678 scope.go:117] "RemoveContainer" containerID="af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.479194 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fd422" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.510918 4678 scope.go:117] "RemoveContainer" containerID="245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.541327 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.548391 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fd422"] Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.566687 4678 scope.go:117] "RemoveContainer" containerID="688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.568306 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2p9wv"] Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.568795 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="registry-server" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.568818 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="registry-server" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.568850 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dacde92-4685-4950-8219-d634cbd0b1cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.568858 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dacde92-4685-4950-8219-d634cbd0b1cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.568871 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="extract-utilities" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.568878 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="extract-utilities" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.568889 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="extract-content" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.568896 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="extract-content" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.569207 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" containerName="registry-server" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.569234 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dacde92-4685-4950-8219-d634cbd0b1cc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.569996 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.572230 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.574913 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.575250 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.575548 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.577437 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2p9wv"] Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.592546 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.592986 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.643677 4678 scope.go:117] "RemoveContainer" containerID="af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.644219 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6\": container with ID starting with af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6 not found: ID does not exist" containerID="af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.644250 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6"} err="failed to get container status \"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6\": rpc error: code = NotFound desc = could not find container \"af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6\": container with ID starting with af0cc08848ef77857230b2bd27c9b515be657e5454fad4e50867070639cea1a6 not found: ID does not exist" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.644269 4678 scope.go:117] "RemoveContainer" containerID="245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.644632 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724\": container with ID starting with 245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724 not found: ID does not exist" containerID="245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.644679 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724"} err="failed to get container status \"245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724\": rpc error: code = NotFound desc = could not find container \"245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724\": container with ID starting with 245d0736c0f57d5c201a058a3ef2420ceb60ee5430c5696be26b9220fa123724 not found: ID does not exist" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.644711 4678 scope.go:117] "RemoveContainer" containerID="688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353" Oct 13 13:13:09 crc kubenswrapper[4678]: E1013 13:13:09.645165 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353\": container with ID starting with 688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353 not found: ID does not exist" containerID="688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.645199 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353"} err="failed to get container status \"688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353\": rpc error: code = NotFound desc = could not find container \"688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353\": container with ID starting with 688674b60c9954cb068ea3c212d0294e4eac6d8c9d8f7db1bee5f4379ae4c353 not found: ID does not exist" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.772250 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5whqx\" (UniqueName: \"kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.772368 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.772954 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.811159 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-844xn" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="registry-server" probeResult="failure" output=< Oct 13 13:13:09 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Oct 13 13:13:09 crc kubenswrapper[4678]: > Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.875167 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.875259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5whqx\" (UniqueName: \"kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.875299 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.879349 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.879414 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.894210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5whqx\" (UniqueName: \"kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx\") pod \"ssh-known-hosts-edpm-deployment-2p9wv\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:09 crc kubenswrapper[4678]: I1013 13:13:09.974340 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:10 crc kubenswrapper[4678]: I1013 13:13:10.508430 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2p9wv"] Oct 13 13:13:10 crc kubenswrapper[4678]: I1013 13:13:10.606461 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601" path="/var/lib/kubelet/pods/e9733c8b-d7a8-4dd6-b6f4-1443ad2c0601/volumes" Oct 13 13:13:11 crc kubenswrapper[4678]: I1013 13:13:11.499526 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" event={"ID":"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e","Type":"ContainerStarted","Data":"1ffa3e11014b915b47ecc75de3c08970fba8a0937a9d72316a3338923f44dc39"} Oct 13 13:13:12 crc kubenswrapper[4678]: I1013 13:13:12.510541 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" event={"ID":"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e","Type":"ContainerStarted","Data":"bd71bef7281129f3671411bb2059b3a116a33895fd40b10bba1e00974cea3875"} Oct 13 13:13:12 crc kubenswrapper[4678]: I1013 13:13:12.553519 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" podStartSLOduration=2.452356476 podStartE2EDuration="3.553500419s" podCreationTimestamp="2025-10-13 13:13:09 +0000 UTC" firstStartedPulling="2025-10-13 13:13:10.5160745 +0000 UTC m=+1778.600612384" lastFinishedPulling="2025-10-13 13:13:11.617218443 +0000 UTC m=+1779.701756327" observedRunningTime="2025-10-13 13:13:12.550942922 +0000 UTC m=+1780.635480826" watchObservedRunningTime="2025-10-13 13:13:12.553500419 +0000 UTC m=+1780.638038293" Oct 13 13:13:18 crc kubenswrapper[4678]: I1013 13:13:18.808539 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:18 crc kubenswrapper[4678]: I1013 13:13:18.871318 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:19 crc kubenswrapper[4678]: I1013 13:13:19.583413 4678 generic.go:334] "Generic (PLEG): container finished" podID="78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" containerID="bd71bef7281129f3671411bb2059b3a116a33895fd40b10bba1e00974cea3875" exitCode=0 Oct 13 13:13:19 crc kubenswrapper[4678]: I1013 13:13:19.583556 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" event={"ID":"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e","Type":"ContainerDied","Data":"bd71bef7281129f3671411bb2059b3a116a33895fd40b10bba1e00974cea3875"} Oct 13 13:13:20 crc kubenswrapper[4678]: I1013 13:13:20.822012 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:13:20 crc kubenswrapper[4678]: I1013 13:13:20.822555 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-844xn" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="registry-server" containerID="cri-o://6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca" gracePeriod=2 Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.003977 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.011285 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5whqx\" (UniqueName: \"kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx\") pod \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.013035 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam\") pod \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.013276 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0\") pod \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\" (UID: \"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.018439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx" (OuterVolumeSpecName: "kube-api-access-5whqx") pod "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" (UID: "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e"). InnerVolumeSpecName "kube-api-access-5whqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.047439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" (UID: "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.058313 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" (UID: "78c20a4a-fd2b-4fa7-bf79-4ae7559b312e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.116963 4678 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.116999 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5whqx\" (UniqueName: \"kubernetes.io/projected/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-kube-api-access-5whqx\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.117014 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78c20a4a-fd2b-4fa7-bf79-4ae7559b312e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.265615 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.319458 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content\") pod \"e5b9f839-bba0-495b-8f42-111f19851bd0\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.319833 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities\") pod \"e5b9f839-bba0-495b-8f42-111f19851bd0\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.319898 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9ddg\" (UniqueName: \"kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg\") pod \"e5b9f839-bba0-495b-8f42-111f19851bd0\" (UID: \"e5b9f839-bba0-495b-8f42-111f19851bd0\") " Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.320756 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities" (OuterVolumeSpecName: "utilities") pod "e5b9f839-bba0-495b-8f42-111f19851bd0" (UID: "e5b9f839-bba0-495b-8f42-111f19851bd0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.324454 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg" (OuterVolumeSpecName: "kube-api-access-j9ddg") pod "e5b9f839-bba0-495b-8f42-111f19851bd0" (UID: "e5b9f839-bba0-495b-8f42-111f19851bd0"). InnerVolumeSpecName "kube-api-access-j9ddg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.405272 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5b9f839-bba0-495b-8f42-111f19851bd0" (UID: "e5b9f839-bba0-495b-8f42-111f19851bd0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.423061 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.423119 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9ddg\" (UniqueName: \"kubernetes.io/projected/e5b9f839-bba0-495b-8f42-111f19851bd0-kube-api-access-j9ddg\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.423133 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5b9f839-bba0-495b-8f42-111f19851bd0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.602214 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" event={"ID":"78c20a4a-fd2b-4fa7-bf79-4ae7559b312e","Type":"ContainerDied","Data":"1ffa3e11014b915b47ecc75de3c08970fba8a0937a9d72316a3338923f44dc39"} Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.602275 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ffa3e11014b915b47ecc75de3c08970fba8a0937a9d72316a3338923f44dc39" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.602230 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2p9wv" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.606576 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerID="6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca" exitCode=0 Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.606617 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerDied","Data":"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca"} Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.606648 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-844xn" event={"ID":"e5b9f839-bba0-495b-8f42-111f19851bd0","Type":"ContainerDied","Data":"11b2970ac500f8fb89d8757c8aa870041f27882b0ed4633f5f84ef38734f6fd2"} Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.606670 4678 scope.go:117] "RemoveContainer" containerID="6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.606817 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-844xn" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.660309 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.679698 4678 scope.go:117] "RemoveContainer" containerID="5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.684531 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-844xn"] Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.706908 4678 scope.go:117] "RemoveContainer" containerID="641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.744264 4678 scope.go:117] "RemoveContainer" containerID="6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.745889 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca\": container with ID starting with 6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca not found: ID does not exist" containerID="6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.746422 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca"} err="failed to get container status \"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca\": rpc error: code = NotFound desc = could not find container \"6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca\": container with ID starting with 6ded12ba306426f5711adc2f0b44207cdcfec96c2b9d5b44a3cbb2deff22fdca not found: ID does not exist" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.746465 4678 scope.go:117] "RemoveContainer" containerID="5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.747030 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb\": container with ID starting with 5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb not found: ID does not exist" containerID="5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.747076 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb"} err="failed to get container status \"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb\": rpc error: code = NotFound desc = could not find container \"5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb\": container with ID starting with 5c01c742e0bca552e79079e35363001e02fa5f7c20de170d0b304c07a9a90bdb not found: ID does not exist" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.747096 4678 scope.go:117] "RemoveContainer" containerID="641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.747365 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1\": container with ID starting with 641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1 not found: ID does not exist" containerID="641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.747413 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1"} err="failed to get container status \"641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1\": rpc error: code = NotFound desc = could not find container \"641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1\": container with ID starting with 641fcc5cb0de63e7a7fdb0dcec88c71b16ee19e1d91d6073444bafda566554b1 not found: ID does not exist" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.752495 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm"] Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.753544 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="extract-content" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.753572 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="extract-content" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.753622 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="extract-utilities" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.753631 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="extract-utilities" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.753686 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="registry-server" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.753696 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="registry-server" Oct 13 13:13:21 crc kubenswrapper[4678]: E1013 13:13:21.753729 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" containerName="ssh-known-hosts-edpm-deployment" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.753737 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" containerName="ssh-known-hosts-edpm-deployment" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.754316 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c20a4a-fd2b-4fa7-bf79-4ae7559b312e" containerName="ssh-known-hosts-edpm-deployment" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.754373 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" containerName="registry-server" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.756226 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.763611 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.763786 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm"] Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.764003 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.764277 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.764421 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.833157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ssq4\" (UniqueName: \"kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.833338 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.833602 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.934498 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.934677 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.934749 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ssq4\" (UniqueName: \"kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.938879 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.939178 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:21 crc kubenswrapper[4678]: I1013 13:13:21.953447 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ssq4\" (UniqueName: \"kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pbbsm\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:22 crc kubenswrapper[4678]: I1013 13:13:22.087029 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:22 crc kubenswrapper[4678]: I1013 13:13:22.611629 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b9f839-bba0-495b-8f42-111f19851bd0" path="/var/lib/kubelet/pods/e5b9f839-bba0-495b-8f42-111f19851bd0/volumes" Oct 13 13:13:22 crc kubenswrapper[4678]: I1013 13:13:22.623765 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm"] Oct 13 13:13:22 crc kubenswrapper[4678]: W1013 13:13:22.631812 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4db4bd12_e81c_4bcc_8f80_ed20807d5856.slice/crio-f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a WatchSource:0}: Error finding container f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a: Status 404 returned error can't find the container with id f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a Oct 13 13:13:23 crc kubenswrapper[4678]: I1013 13:13:23.637210 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" event={"ID":"4db4bd12-e81c-4bcc-8f80-ed20807d5856","Type":"ContainerStarted","Data":"47eac08f260e84e1a45cd034c488dccefb584530d76a757ae54879e0d930e0e3"} Oct 13 13:13:23 crc kubenswrapper[4678]: I1013 13:13:23.637563 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" event={"ID":"4db4bd12-e81c-4bcc-8f80-ed20807d5856","Type":"ContainerStarted","Data":"f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a"} Oct 13 13:13:23 crc kubenswrapper[4678]: I1013 13:13:23.653485 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" podStartSLOduration=1.92221913 podStartE2EDuration="2.653466954s" podCreationTimestamp="2025-10-13 13:13:21 +0000 UTC" firstStartedPulling="2025-10-13 13:13:22.63353845 +0000 UTC m=+1790.718076324" lastFinishedPulling="2025-10-13 13:13:23.364786254 +0000 UTC m=+1791.449324148" observedRunningTime="2025-10-13 13:13:23.651659767 +0000 UTC m=+1791.736197671" watchObservedRunningTime="2025-10-13 13:13:23.653466954 +0000 UTC m=+1791.738004838" Oct 13 13:13:24 crc kubenswrapper[4678]: I1013 13:13:24.592603 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:13:24 crc kubenswrapper[4678]: E1013 13:13:24.593218 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:13:32 crc kubenswrapper[4678]: I1013 13:13:32.728241 4678 generic.go:334] "Generic (PLEG): container finished" podID="4db4bd12-e81c-4bcc-8f80-ed20807d5856" containerID="47eac08f260e84e1a45cd034c488dccefb584530d76a757ae54879e0d930e0e3" exitCode=0 Oct 13 13:13:32 crc kubenswrapper[4678]: I1013 13:13:32.728329 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" event={"ID":"4db4bd12-e81c-4bcc-8f80-ed20807d5856","Type":"ContainerDied","Data":"47eac08f260e84e1a45cd034c488dccefb584530d76a757ae54879e0d930e0e3"} Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.166917 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.255681 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key\") pod \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.255893 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory\") pod \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.255923 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ssq4\" (UniqueName: \"kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4\") pod \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\" (UID: \"4db4bd12-e81c-4bcc-8f80-ed20807d5856\") " Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.261687 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4" (OuterVolumeSpecName: "kube-api-access-9ssq4") pod "4db4bd12-e81c-4bcc-8f80-ed20807d5856" (UID: "4db4bd12-e81c-4bcc-8f80-ed20807d5856"). InnerVolumeSpecName "kube-api-access-9ssq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.289004 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory" (OuterVolumeSpecName: "inventory") pod "4db4bd12-e81c-4bcc-8f80-ed20807d5856" (UID: "4db4bd12-e81c-4bcc-8f80-ed20807d5856"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.289284 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4db4bd12-e81c-4bcc-8f80-ed20807d5856" (UID: "4db4bd12-e81c-4bcc-8f80-ed20807d5856"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.358694 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.358734 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ssq4\" (UniqueName: \"kubernetes.io/projected/4db4bd12-e81c-4bcc-8f80-ed20807d5856-kube-api-access-9ssq4\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.358745 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4db4bd12-e81c-4bcc-8f80-ed20807d5856-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.750227 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" event={"ID":"4db4bd12-e81c-4bcc-8f80-ed20807d5856","Type":"ContainerDied","Data":"f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a"} Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.750273 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f953cacd428ca16b6d0025573cf57087544abf9181add174806767a79f0e2a5a" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.750342 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pbbsm" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.817604 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv"] Oct 13 13:13:34 crc kubenswrapper[4678]: E1013 13:13:34.820358 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db4bd12-e81c-4bcc-8f80-ed20807d5856" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.820410 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db4bd12-e81c-4bcc-8f80-ed20807d5856" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.820660 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db4bd12-e81c-4bcc-8f80-ed20807d5856" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.821474 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.826036 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.826081 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.826122 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.826256 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.831037 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv"] Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.969127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57tmw\" (UniqueName: \"kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.969184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:34 crc kubenswrapper[4678]: I1013 13:13:34.969207 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.070727 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57tmw\" (UniqueName: \"kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.070848 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.070891 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.077666 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.077997 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.094968 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57tmw\" (UniqueName: \"kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.153611 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.660880 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv"] Oct 13 13:13:35 crc kubenswrapper[4678]: I1013 13:13:35.759742 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" event={"ID":"b8dc18d2-df42-4546-981a-cfe13dcd879f","Type":"ContainerStarted","Data":"5815a153b14dac7884433247f76b18c5b25963ca4e842945c735090980606696"} Oct 13 13:13:36 crc kubenswrapper[4678]: I1013 13:13:36.592836 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:13:36 crc kubenswrapper[4678]: I1013 13:13:36.770394 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" event={"ID":"b8dc18d2-df42-4546-981a-cfe13dcd879f","Type":"ContainerStarted","Data":"eff1c167400728f66e95a16dc3894b80bd99053ac7b2740bdb935b7e3e505ca9"} Oct 13 13:13:36 crc kubenswrapper[4678]: I1013 13:13:36.799939 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" podStartSLOduration=2.367356639 podStartE2EDuration="2.799916263s" podCreationTimestamp="2025-10-13 13:13:34 +0000 UTC" firstStartedPulling="2025-10-13 13:13:35.667098278 +0000 UTC m=+1803.751636172" lastFinishedPulling="2025-10-13 13:13:36.099657912 +0000 UTC m=+1804.184195796" observedRunningTime="2025-10-13 13:13:36.785994502 +0000 UTC m=+1804.870532386" watchObservedRunningTime="2025-10-13 13:13:36.799916263 +0000 UTC m=+1804.884454147" Oct 13 13:13:37 crc kubenswrapper[4678]: I1013 13:13:37.784105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415"} Oct 13 13:13:45 crc kubenswrapper[4678]: I1013 13:13:45.858801 4678 generic.go:334] "Generic (PLEG): container finished" podID="b8dc18d2-df42-4546-981a-cfe13dcd879f" containerID="eff1c167400728f66e95a16dc3894b80bd99053ac7b2740bdb935b7e3e505ca9" exitCode=0 Oct 13 13:13:45 crc kubenswrapper[4678]: I1013 13:13:45.858873 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" event={"ID":"b8dc18d2-df42-4546-981a-cfe13dcd879f","Type":"ContainerDied","Data":"eff1c167400728f66e95a16dc3894b80bd99053ac7b2740bdb935b7e3e505ca9"} Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.343635 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.439564 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57tmw\" (UniqueName: \"kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw\") pod \"b8dc18d2-df42-4546-981a-cfe13dcd879f\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.439692 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key\") pod \"b8dc18d2-df42-4546-981a-cfe13dcd879f\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.439738 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory\") pod \"b8dc18d2-df42-4546-981a-cfe13dcd879f\" (UID: \"b8dc18d2-df42-4546-981a-cfe13dcd879f\") " Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.445884 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw" (OuterVolumeSpecName: "kube-api-access-57tmw") pod "b8dc18d2-df42-4546-981a-cfe13dcd879f" (UID: "b8dc18d2-df42-4546-981a-cfe13dcd879f"). InnerVolumeSpecName "kube-api-access-57tmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.468502 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8dc18d2-df42-4546-981a-cfe13dcd879f" (UID: "b8dc18d2-df42-4546-981a-cfe13dcd879f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.473336 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory" (OuterVolumeSpecName: "inventory") pod "b8dc18d2-df42-4546-981a-cfe13dcd879f" (UID: "b8dc18d2-df42-4546-981a-cfe13dcd879f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.542438 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.542476 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8dc18d2-df42-4546-981a-cfe13dcd879f-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.542486 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57tmw\" (UniqueName: \"kubernetes.io/projected/b8dc18d2-df42-4546-981a-cfe13dcd879f-kube-api-access-57tmw\") on node \"crc\" DevicePath \"\"" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.877479 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" event={"ID":"b8dc18d2-df42-4546-981a-cfe13dcd879f","Type":"ContainerDied","Data":"5815a153b14dac7884433247f76b18c5b25963ca4e842945c735090980606696"} Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.877793 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5815a153b14dac7884433247f76b18c5b25963ca4e842945c735090980606696" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.877860 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.991251 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6"] Oct 13 13:13:47 crc kubenswrapper[4678]: E1013 13:13:47.991658 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8dc18d2-df42-4546-981a-cfe13dcd879f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.991676 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8dc18d2-df42-4546-981a-cfe13dcd879f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.991863 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8dc18d2-df42-4546-981a-cfe13dcd879f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.992509 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.998682 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.999367 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.999750 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:13:47 crc kubenswrapper[4678]: I1013 13:13:47.999970 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.000232 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.000599 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.000807 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.001461 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.010772 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6"] Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154009 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154080 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154147 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154172 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154299 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154491 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154534 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154566 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkqls\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154615 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154770 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154854 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.154941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.155132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.155182 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.257117 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.257189 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkqls\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258125 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258170 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258204 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258253 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258331 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258355 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258394 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258475 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258498 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.258573 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.263162 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.263435 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.264104 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.264281 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265220 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265518 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265714 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265855 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265916 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.265979 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.267356 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.267918 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.271323 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.280423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkqls\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.353944 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:13:48 crc kubenswrapper[4678]: I1013 13:13:48.897669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6"] Oct 13 13:13:49 crc kubenswrapper[4678]: I1013 13:13:49.901715 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" event={"ID":"3608a541-f8d9-4931-a7d1-cbf90f6c505f","Type":"ContainerStarted","Data":"08414654e1f86a8e4b11b8dbf636ce37b6117a6f3cdc3216d570942a2d84951d"} Oct 13 13:13:49 crc kubenswrapper[4678]: I1013 13:13:49.902930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" event={"ID":"3608a541-f8d9-4931-a7d1-cbf90f6c505f","Type":"ContainerStarted","Data":"4ff2962fe86ac0b3550a2f9d906628b0809d3ecac37f235d2ef7088a17f3a61b"} Oct 13 13:13:49 crc kubenswrapper[4678]: I1013 13:13:49.933168 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" podStartSLOduration=2.354684297 podStartE2EDuration="2.933126007s" podCreationTimestamp="2025-10-13 13:13:47 +0000 UTC" firstStartedPulling="2025-10-13 13:13:48.906666362 +0000 UTC m=+1816.991204246" lastFinishedPulling="2025-10-13 13:13:49.485108032 +0000 UTC m=+1817.569645956" observedRunningTime="2025-10-13 13:13:49.926728381 +0000 UTC m=+1818.011266285" watchObservedRunningTime="2025-10-13 13:13:49.933126007 +0000 UTC m=+1818.017663891" Oct 13 13:14:31 crc kubenswrapper[4678]: I1013 13:14:31.274349 4678 generic.go:334] "Generic (PLEG): container finished" podID="3608a541-f8d9-4931-a7d1-cbf90f6c505f" containerID="08414654e1f86a8e4b11b8dbf636ce37b6117a6f3cdc3216d570942a2d84951d" exitCode=0 Oct 13 13:14:31 crc kubenswrapper[4678]: I1013 13:14:31.274445 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" event={"ID":"3608a541-f8d9-4931-a7d1-cbf90f6c505f","Type":"ContainerDied","Data":"08414654e1f86a8e4b11b8dbf636ce37b6117a6f3cdc3216d570942a2d84951d"} Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.773268 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838586 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838628 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838656 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838705 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838731 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838782 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838799 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkqls\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838854 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838881 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838902 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838933 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.838961 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.839096 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.839151 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle\") pod \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\" (UID: \"3608a541-f8d9-4931-a7d1-cbf90f6c505f\") " Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.845099 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.845536 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.846091 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.846747 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.847728 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.848373 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.849205 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.849213 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.850388 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.851472 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.854018 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.854570 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls" (OuterVolumeSpecName: "kube-api-access-qkqls") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "kube-api-access-qkqls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.875419 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory" (OuterVolumeSpecName: "inventory") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.881223 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3608a541-f8d9-4931-a7d1-cbf90f6c505f" (UID: "3608a541-f8d9-4931-a7d1-cbf90f6c505f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941876 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941915 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941947 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941961 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941970 4678 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941979 4678 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941987 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.941995 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942004 4678 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942031 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942042 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942066 4678 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942076 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3608a541-f8d9-4931-a7d1-cbf90f6c505f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:32 crc kubenswrapper[4678]: I1013 13:14:32.942084 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkqls\" (UniqueName: \"kubernetes.io/projected/3608a541-f8d9-4931-a7d1-cbf90f6c505f-kube-api-access-qkqls\") on node \"crc\" DevicePath \"\"" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.296309 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" event={"ID":"3608a541-f8d9-4931-a7d1-cbf90f6c505f","Type":"ContainerDied","Data":"4ff2962fe86ac0b3550a2f9d906628b0809d3ecac37f235d2ef7088a17f3a61b"} Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.296571 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.296583 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff2962fe86ac0b3550a2f9d906628b0809d3ecac37f235d2ef7088a17f3a61b" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.449723 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85"] Oct 13 13:14:33 crc kubenswrapper[4678]: E1013 13:14:33.450179 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3608a541-f8d9-4931-a7d1-cbf90f6c505f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.450197 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3608a541-f8d9-4931-a7d1-cbf90f6c505f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.450391 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3608a541-f8d9-4931-a7d1-cbf90f6c505f" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.451017 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.453082 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.453191 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.453528 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.453598 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.454413 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.462099 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85"] Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.557481 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.557798 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhsnt\" (UniqueName: \"kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.557877 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.558088 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.558293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.659952 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.660045 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.660129 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.660168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.660197 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhsnt\" (UniqueName: \"kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.662198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.665854 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.666958 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.671371 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.690760 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhsnt\" (UniqueName: \"kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lvc85\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:33 crc kubenswrapper[4678]: I1013 13:14:33.767313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:14:34 crc kubenswrapper[4678]: I1013 13:14:34.283894 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85"] Oct 13 13:14:34 crc kubenswrapper[4678]: I1013 13:14:34.291314 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:14:34 crc kubenswrapper[4678]: I1013 13:14:34.307067 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" event={"ID":"416f5c49-3fd6-4789-924a-4e0ba0279fa5","Type":"ContainerStarted","Data":"c1df36bf2a5dab77f879ceea7a04a3382ee05a98c3b01535fde1e70707ce9af4"} Oct 13 13:14:35 crc kubenswrapper[4678]: I1013 13:14:35.316092 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" event={"ID":"416f5c49-3fd6-4789-924a-4e0ba0279fa5","Type":"ContainerStarted","Data":"0bc3b7e37b8e27a331499eff6db61f35d2a9f986d46ffd660bf86dcebebf0682"} Oct 13 13:14:35 crc kubenswrapper[4678]: I1013 13:14:35.334184 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" podStartSLOduration=1.645162869 podStartE2EDuration="2.334167248s" podCreationTimestamp="2025-10-13 13:14:33 +0000 UTC" firstStartedPulling="2025-10-13 13:14:34.291134923 +0000 UTC m=+1862.375672797" lastFinishedPulling="2025-10-13 13:14:34.980139292 +0000 UTC m=+1863.064677176" observedRunningTime="2025-10-13 13:14:35.332453903 +0000 UTC m=+1863.416991797" watchObservedRunningTime="2025-10-13 13:14:35.334167248 +0000 UTC m=+1863.418705132" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.188052 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt"] Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.189836 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.192617 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.197388 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.197672 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt"] Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.274617 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.274663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86cq8\" (UniqueName: \"kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.274685 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.376574 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.376621 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86cq8\" (UniqueName: \"kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.376654 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.377936 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.384911 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.393233 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86cq8\" (UniqueName: \"kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8\") pod \"collect-profiles-29339355-vrgqt\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.517345 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:00 crc kubenswrapper[4678]: I1013 13:15:00.982768 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt"] Oct 13 13:15:01 crc kubenswrapper[4678]: I1013 13:15:01.571510 4678 generic.go:334] "Generic (PLEG): container finished" podID="733935b6-03ec-46d7-9a10-45b738d4b071" containerID="7e45c54e4e0c0f203a6ca84acc99c46411920a618e1768076dbf614a10ae21f2" exitCode=0 Oct 13 13:15:01 crc kubenswrapper[4678]: I1013 13:15:01.571555 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" event={"ID":"733935b6-03ec-46d7-9a10-45b738d4b071","Type":"ContainerDied","Data":"7e45c54e4e0c0f203a6ca84acc99c46411920a618e1768076dbf614a10ae21f2"} Oct 13 13:15:01 crc kubenswrapper[4678]: I1013 13:15:01.571823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" event={"ID":"733935b6-03ec-46d7-9a10-45b738d4b071","Type":"ContainerStarted","Data":"25704d497adf402ed77bb49c26ba5553f263f541e1455971d124e892eb1731a6"} Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.908650 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.931016 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86cq8\" (UniqueName: \"kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8\") pod \"733935b6-03ec-46d7-9a10-45b738d4b071\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.931095 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume\") pod \"733935b6-03ec-46d7-9a10-45b738d4b071\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.931430 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume\") pod \"733935b6-03ec-46d7-9a10-45b738d4b071\" (UID: \"733935b6-03ec-46d7-9a10-45b738d4b071\") " Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.931976 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume" (OuterVolumeSpecName: "config-volume") pod "733935b6-03ec-46d7-9a10-45b738d4b071" (UID: "733935b6-03ec-46d7-9a10-45b738d4b071"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.937672 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8" (OuterVolumeSpecName: "kube-api-access-86cq8") pod "733935b6-03ec-46d7-9a10-45b738d4b071" (UID: "733935b6-03ec-46d7-9a10-45b738d4b071"). InnerVolumeSpecName "kube-api-access-86cq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:15:02 crc kubenswrapper[4678]: I1013 13:15:02.938377 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "733935b6-03ec-46d7-9a10-45b738d4b071" (UID: "733935b6-03ec-46d7-9a10-45b738d4b071"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.033419 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/733935b6-03ec-46d7-9a10-45b738d4b071-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.033449 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86cq8\" (UniqueName: \"kubernetes.io/projected/733935b6-03ec-46d7-9a10-45b738d4b071-kube-api-access-86cq8\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.033459 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/733935b6-03ec-46d7-9a10-45b738d4b071-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.592810 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" event={"ID":"733935b6-03ec-46d7-9a10-45b738d4b071","Type":"ContainerDied","Data":"25704d497adf402ed77bb49c26ba5553f263f541e1455971d124e892eb1731a6"} Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.592839 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25704d497adf402ed77bb49c26ba5553f263f541e1455971d124e892eb1731a6" Oct 13 13:15:03 crc kubenswrapper[4678]: I1013 13:15:03.592885 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339355-vrgqt" Oct 13 13:15:39 crc kubenswrapper[4678]: I1013 13:15:39.959937 4678 generic.go:334] "Generic (PLEG): container finished" podID="416f5c49-3fd6-4789-924a-4e0ba0279fa5" containerID="0bc3b7e37b8e27a331499eff6db61f35d2a9f986d46ffd660bf86dcebebf0682" exitCode=0 Oct 13 13:15:39 crc kubenswrapper[4678]: I1013 13:15:39.960033 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" event={"ID":"416f5c49-3fd6-4789-924a-4e0ba0279fa5","Type":"ContainerDied","Data":"0bc3b7e37b8e27a331499eff6db61f35d2a9f986d46ffd660bf86dcebebf0682"} Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.406899 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.519775 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhsnt\" (UniqueName: \"kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt\") pod \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.520176 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle\") pod \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.520257 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key\") pod \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.520321 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory\") pod \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.520389 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0\") pod \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\" (UID: \"416f5c49-3fd6-4789-924a-4e0ba0279fa5\") " Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.526087 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "416f5c49-3fd6-4789-924a-4e0ba0279fa5" (UID: "416f5c49-3fd6-4789-924a-4e0ba0279fa5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.527385 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt" (OuterVolumeSpecName: "kube-api-access-jhsnt") pod "416f5c49-3fd6-4789-924a-4e0ba0279fa5" (UID: "416f5c49-3fd6-4789-924a-4e0ba0279fa5"). InnerVolumeSpecName "kube-api-access-jhsnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.545338 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "416f5c49-3fd6-4789-924a-4e0ba0279fa5" (UID: "416f5c49-3fd6-4789-924a-4e0ba0279fa5"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.549561 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "416f5c49-3fd6-4789-924a-4e0ba0279fa5" (UID: "416f5c49-3fd6-4789-924a-4e0ba0279fa5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.553906 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory" (OuterVolumeSpecName: "inventory") pod "416f5c49-3fd6-4789-924a-4e0ba0279fa5" (UID: "416f5c49-3fd6-4789-924a-4e0ba0279fa5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.623103 4678 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.623146 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhsnt\" (UniqueName: \"kubernetes.io/projected/416f5c49-3fd6-4789-924a-4e0ba0279fa5-kube-api-access-jhsnt\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.623158 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.623170 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.623181 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/416f5c49-3fd6-4789-924a-4e0ba0279fa5-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.983750 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" event={"ID":"416f5c49-3fd6-4789-924a-4e0ba0279fa5","Type":"ContainerDied","Data":"c1df36bf2a5dab77f879ceea7a04a3382ee05a98c3b01535fde1e70707ce9af4"} Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.983793 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1df36bf2a5dab77f879ceea7a04a3382ee05a98c3b01535fde1e70707ce9af4" Oct 13 13:15:41 crc kubenswrapper[4678]: I1013 13:15:41.983867 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lvc85" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.087221 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6"] Oct 13 13:15:42 crc kubenswrapper[4678]: E1013 13:15:42.087862 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416f5c49-3fd6-4789-924a-4e0ba0279fa5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.087894 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="416f5c49-3fd6-4789-924a-4e0ba0279fa5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 13:15:42 crc kubenswrapper[4678]: E1013 13:15:42.087943 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733935b6-03ec-46d7-9a10-45b738d4b071" containerName="collect-profiles" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.087962 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="733935b6-03ec-46d7-9a10-45b738d4b071" containerName="collect-profiles" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.088372 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="416f5c49-3fd6-4789-924a-4e0ba0279fa5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.088419 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="733935b6-03ec-46d7-9a10-45b738d4b071" containerName="collect-profiles" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.089456 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.092514 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.092920 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.093453 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.093536 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.093837 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.093913 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.116240 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6"] Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.235472 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.235551 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.235708 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.235811 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmqpl\" (UniqueName: \"kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.235926 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.236191 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.337768 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.337819 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.337860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.337903 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmqpl\" (UniqueName: \"kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.337988 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.338090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.342330 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.342441 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.344325 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.344597 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.346551 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.356976 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmqpl\" (UniqueName: \"kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.425825 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.933410 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6"] Oct 13 13:15:42 crc kubenswrapper[4678]: W1013 13:15:42.933725 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a30e130_e610_4241_beb7_0a5d88fda83b.slice/crio-16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d WatchSource:0}: Error finding container 16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d: Status 404 returned error can't find the container with id 16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d Oct 13 13:15:42 crc kubenswrapper[4678]: I1013 13:15:42.991858 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" event={"ID":"8a30e130-e610-4241-beb7-0a5d88fda83b","Type":"ContainerStarted","Data":"16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d"} Oct 13 13:15:44 crc kubenswrapper[4678]: I1013 13:15:44.001786 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" event={"ID":"8a30e130-e610-4241-beb7-0a5d88fda83b","Type":"ContainerStarted","Data":"c4ae4acf787454581f02f6a8474b55df00ef5dd0195d05a6fd4194538d61e6c8"} Oct 13 13:15:44 crc kubenswrapper[4678]: I1013 13:15:44.029185 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" podStartSLOduration=1.640758551 podStartE2EDuration="2.029130947s" podCreationTimestamp="2025-10-13 13:15:42 +0000 UTC" firstStartedPulling="2025-10-13 13:15:42.936611359 +0000 UTC m=+1931.021149243" lastFinishedPulling="2025-10-13 13:15:43.324983765 +0000 UTC m=+1931.409521639" observedRunningTime="2025-10-13 13:15:44.017658259 +0000 UTC m=+1932.102196153" watchObservedRunningTime="2025-10-13 13:15:44.029130947 +0000 UTC m=+1932.113668841" Oct 13 13:15:55 crc kubenswrapper[4678]: I1013 13:15:55.506764 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:15:55 crc kubenswrapper[4678]: I1013 13:15:55.507461 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:16:25 crc kubenswrapper[4678]: I1013 13:16:25.505846 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:16:25 crc kubenswrapper[4678]: I1013 13:16:25.506522 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:16:33 crc kubenswrapper[4678]: I1013 13:16:33.403281 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a30e130-e610-4241-beb7-0a5d88fda83b" containerID="c4ae4acf787454581f02f6a8474b55df00ef5dd0195d05a6fd4194538d61e6c8" exitCode=0 Oct 13 13:16:33 crc kubenswrapper[4678]: I1013 13:16:33.403347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" event={"ID":"8a30e130-e610-4241-beb7-0a5d88fda83b","Type":"ContainerDied","Data":"c4ae4acf787454581f02f6a8474b55df00ef5dd0195d05a6fd4194538d61e6c8"} Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.818756 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947068 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947218 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947274 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmqpl\" (UniqueName: \"kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947311 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947363 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.947442 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory\") pod \"8a30e130-e610-4241-beb7-0a5d88fda83b\" (UID: \"8a30e130-e610-4241-beb7-0a5d88fda83b\") " Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.952750 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.955463 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl" (OuterVolumeSpecName: "kube-api-access-mmqpl") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "kube-api-access-mmqpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.975732 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.979820 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.980450 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:16:34 crc kubenswrapper[4678]: I1013 13:16:34.981864 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory" (OuterVolumeSpecName: "inventory") pod "8a30e130-e610-4241-beb7-0a5d88fda83b" (UID: "8a30e130-e610-4241-beb7-0a5d88fda83b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049352 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049395 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049410 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmqpl\" (UniqueName: \"kubernetes.io/projected/8a30e130-e610-4241-beb7-0a5d88fda83b-kube-api-access-mmqpl\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049422 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049432 4678 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.049443 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8a30e130-e610-4241-beb7-0a5d88fda83b-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.434480 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" event={"ID":"8a30e130-e610-4241-beb7-0a5d88fda83b","Type":"ContainerDied","Data":"16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d"} Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.434519 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16447477c973e989e1610e5a7ddb2a8abc6d4716069f2148c543ddf08747177d" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.434538 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.547829 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6"] Oct 13 13:16:35 crc kubenswrapper[4678]: E1013 13:16:35.548476 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a30e130-e610-4241-beb7-0a5d88fda83b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.548501 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a30e130-e610-4241-beb7-0a5d88fda83b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.548751 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a30e130-e610-4241-beb7-0a5d88fda83b" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.549625 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.552612 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.552660 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.552664 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.552711 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.554095 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.557701 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6"] Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.659698 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.659814 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.660139 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.660192 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.660321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jl29\" (UniqueName: \"kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.761909 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.761979 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.762135 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jl29\" (UniqueName: \"kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.762266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.762346 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.766318 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.766821 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.769536 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.771038 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.781185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jl29\" (UniqueName: \"kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-m62j6\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:35 crc kubenswrapper[4678]: I1013 13:16:35.870676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:16:36 crc kubenswrapper[4678]: I1013 13:16:36.374136 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6"] Oct 13 13:16:36 crc kubenswrapper[4678]: I1013 13:16:36.448433 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" event={"ID":"08cf88a5-76f1-455d-a838-9d8e37902c6e","Type":"ContainerStarted","Data":"cf1c55340ab248f4d8f0181927251ef82b4bd6380e557e38cc9eec7d178a07eb"} Oct 13 13:16:37 crc kubenswrapper[4678]: I1013 13:16:37.457747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" event={"ID":"08cf88a5-76f1-455d-a838-9d8e37902c6e","Type":"ContainerStarted","Data":"75d35c54549e9e0d14023656a4fe037798b03719e074ef0739b9baed70f17200"} Oct 13 13:16:37 crc kubenswrapper[4678]: I1013 13:16:37.472220 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" podStartSLOduration=1.684924046 podStartE2EDuration="2.472198561s" podCreationTimestamp="2025-10-13 13:16:35 +0000 UTC" firstStartedPulling="2025-10-13 13:16:36.380724891 +0000 UTC m=+1984.465262775" lastFinishedPulling="2025-10-13 13:16:37.167999406 +0000 UTC m=+1985.252537290" observedRunningTime="2025-10-13 13:16:37.47061916 +0000 UTC m=+1985.555157044" watchObservedRunningTime="2025-10-13 13:16:37.472198561 +0000 UTC m=+1985.556736445" Oct 13 13:16:55 crc kubenswrapper[4678]: I1013 13:16:55.505839 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:16:55 crc kubenswrapper[4678]: I1013 13:16:55.506579 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:16:55 crc kubenswrapper[4678]: I1013 13:16:55.506639 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:16:55 crc kubenswrapper[4678]: I1013 13:16:55.507724 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:16:55 crc kubenswrapper[4678]: I1013 13:16:55.507809 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415" gracePeriod=600 Oct 13 13:16:56 crc kubenswrapper[4678]: I1013 13:16:56.640462 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415" exitCode=0 Oct 13 13:16:56 crc kubenswrapper[4678]: I1013 13:16:56.640559 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415"} Oct 13 13:16:56 crc kubenswrapper[4678]: I1013 13:16:56.642010 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d"} Oct 13 13:16:56 crc kubenswrapper[4678]: I1013 13:16:56.642110 4678 scope.go:117] "RemoveContainer" containerID="277810e93d3b025398aab7f22d4e3d1ce2f2a2accc497b0fd057c0301af25e98" Oct 13 13:18:55 crc kubenswrapper[4678]: I1013 13:18:55.505670 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:18:55 crc kubenswrapper[4678]: I1013 13:18:55.506260 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.742449 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z84r9"] Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.746284 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.753296 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z84r9"] Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.932689 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-utilities\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.932778 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxwsg\" (UniqueName: \"kubernetes.io/projected/cb08d1b2-77d0-45bf-9a67-3b32448063a1-kube-api-access-rxwsg\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:16 crc kubenswrapper[4678]: I1013 13:19:16.932882 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-catalog-content\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.034324 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxwsg\" (UniqueName: \"kubernetes.io/projected/cb08d1b2-77d0-45bf-9a67-3b32448063a1-kube-api-access-rxwsg\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.034435 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-catalog-content\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.034606 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-utilities\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.035018 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-catalog-content\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.035126 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb08d1b2-77d0-45bf-9a67-3b32448063a1-utilities\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.061017 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxwsg\" (UniqueName: \"kubernetes.io/projected/cb08d1b2-77d0-45bf-9a67-3b32448063a1-kube-api-access-rxwsg\") pod \"community-operators-z84r9\" (UID: \"cb08d1b2-77d0-45bf-9a67-3b32448063a1\") " pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.069523 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.684264 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z84r9"] Oct 13 13:19:17 crc kubenswrapper[4678]: I1013 13:19:17.886849 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z84r9" event={"ID":"cb08d1b2-77d0-45bf-9a67-3b32448063a1","Type":"ContainerStarted","Data":"df2d484ed956555819844435602a7576dcc09d6b229f51c997145984399bc1c2"} Oct 13 13:19:18 crc kubenswrapper[4678]: I1013 13:19:18.897635 4678 generic.go:334] "Generic (PLEG): container finished" podID="cb08d1b2-77d0-45bf-9a67-3b32448063a1" containerID="7b6df7ddf9b77e664ebec9ed57de84bed7a5bb66b648e405947bb6f5adb343a3" exitCode=0 Oct 13 13:19:18 crc kubenswrapper[4678]: I1013 13:19:18.897720 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z84r9" event={"ID":"cb08d1b2-77d0-45bf-9a67-3b32448063a1","Type":"ContainerDied","Data":"7b6df7ddf9b77e664ebec9ed57de84bed7a5bb66b648e405947bb6f5adb343a3"} Oct 13 13:19:25 crc kubenswrapper[4678]: I1013 13:19:25.505742 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:19:25 crc kubenswrapper[4678]: I1013 13:19:25.506532 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:19:25 crc kubenswrapper[4678]: I1013 13:19:25.967938 4678 generic.go:334] "Generic (PLEG): container finished" podID="cb08d1b2-77d0-45bf-9a67-3b32448063a1" containerID="5cba6f578bbf0607a208fa2230ba9b403ddaea26d2a3d3904757e39b6e5c854a" exitCode=0 Oct 13 13:19:25 crc kubenswrapper[4678]: I1013 13:19:25.968011 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z84r9" event={"ID":"cb08d1b2-77d0-45bf-9a67-3b32448063a1","Type":"ContainerDied","Data":"5cba6f578bbf0607a208fa2230ba9b403ddaea26d2a3d3904757e39b6e5c854a"} Oct 13 13:19:27 crc kubenswrapper[4678]: I1013 13:19:27.984920 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z84r9" event={"ID":"cb08d1b2-77d0-45bf-9a67-3b32448063a1","Type":"ContainerStarted","Data":"4fc3c996f43688929c2c38c944a37740c7f06eeffa7ff80112d9d769c5bf28f0"} Oct 13 13:19:28 crc kubenswrapper[4678]: I1013 13:19:28.006513 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z84r9" podStartSLOduration=3.928304201 podStartE2EDuration="12.006496373s" podCreationTimestamp="2025-10-13 13:19:16 +0000 UTC" firstStartedPulling="2025-10-13 13:19:18.900345012 +0000 UTC m=+2146.984882896" lastFinishedPulling="2025-10-13 13:19:26.978537184 +0000 UTC m=+2155.063075068" observedRunningTime="2025-10-13 13:19:28.003734151 +0000 UTC m=+2156.088272055" watchObservedRunningTime="2025-10-13 13:19:28.006496373 +0000 UTC m=+2156.091034257" Oct 13 13:19:37 crc kubenswrapper[4678]: I1013 13:19:37.071474 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:37 crc kubenswrapper[4678]: I1013 13:19:37.072087 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:37 crc kubenswrapper[4678]: I1013 13:19:37.121628 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.118251 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z84r9" Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.182721 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z84r9"] Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.222209 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.222509 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hkpp7" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="registry-server" containerID="cri-o://03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc" gracePeriod=2 Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.755322 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.924526 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities\") pod \"f4be8bfa-2d9f-490b-b239-852793e85b6b\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.924600 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content\") pod \"f4be8bfa-2d9f-490b-b239-852793e85b6b\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.924641 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq6qc\" (UniqueName: \"kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc\") pod \"f4be8bfa-2d9f-490b-b239-852793e85b6b\" (UID: \"f4be8bfa-2d9f-490b-b239-852793e85b6b\") " Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.927367 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities" (OuterVolumeSpecName: "utilities") pod "f4be8bfa-2d9f-490b-b239-852793e85b6b" (UID: "f4be8bfa-2d9f-490b-b239-852793e85b6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:19:38 crc kubenswrapper[4678]: I1013 13:19:38.938649 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc" (OuterVolumeSpecName: "kube-api-access-rq6qc") pod "f4be8bfa-2d9f-490b-b239-852793e85b6b" (UID: "f4be8bfa-2d9f-490b-b239-852793e85b6b"). InnerVolumeSpecName "kube-api-access-rq6qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.016363 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4be8bfa-2d9f-490b-b239-852793e85b6b" (UID: "f4be8bfa-2d9f-490b-b239-852793e85b6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.026683 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.026725 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4be8bfa-2d9f-490b-b239-852793e85b6b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.026738 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq6qc\" (UniqueName: \"kubernetes.io/projected/f4be8bfa-2d9f-490b-b239-852793e85b6b-kube-api-access-rq6qc\") on node \"crc\" DevicePath \"\"" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.083747 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerID="03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc" exitCode=0 Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.083969 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerDied","Data":"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc"} Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.084239 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkpp7" event={"ID":"f4be8bfa-2d9f-490b-b239-852793e85b6b","Type":"ContainerDied","Data":"cb5975440ede515e9c46bce351a488d6ef761fbd56e24cc97cc754a392b283a4"} Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.084262 4678 scope.go:117] "RemoveContainer" containerID="03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.084109 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkpp7" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.119232 4678 scope.go:117] "RemoveContainer" containerID="75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.123618 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.135597 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hkpp7"] Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.164631 4678 scope.go:117] "RemoveContainer" containerID="b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.189264 4678 scope.go:117] "RemoveContainer" containerID="03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc" Oct 13 13:19:39 crc kubenswrapper[4678]: E1013 13:19:39.196677 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc\": container with ID starting with 03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc not found: ID does not exist" containerID="03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.196732 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc"} err="failed to get container status \"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc\": rpc error: code = NotFound desc = could not find container \"03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc\": container with ID starting with 03fcb575465523d3c28bf078e8ff4392e8738ccbdad8524a6af2ddfc0beb77bc not found: ID does not exist" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.196765 4678 scope.go:117] "RemoveContainer" containerID="75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f" Oct 13 13:19:39 crc kubenswrapper[4678]: E1013 13:19:39.197274 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f\": container with ID starting with 75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f not found: ID does not exist" containerID="75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.197347 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f"} err="failed to get container status \"75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f\": rpc error: code = NotFound desc = could not find container \"75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f\": container with ID starting with 75d161d07ae6db64526e6e201b80bd21acee395dbbb7bc6bde653a3dc391ac7f not found: ID does not exist" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.197397 4678 scope.go:117] "RemoveContainer" containerID="b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5" Oct 13 13:19:39 crc kubenswrapper[4678]: E1013 13:19:39.197738 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5\": container with ID starting with b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5 not found: ID does not exist" containerID="b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5" Oct 13 13:19:39 crc kubenswrapper[4678]: I1013 13:19:39.197774 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5"} err="failed to get container status \"b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5\": rpc error: code = NotFound desc = could not find container \"b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5\": container with ID starting with b353f5177e322e71e17897ce1351dba9a40b63a241994d35ac4bef3d5c1cfde5 not found: ID does not exist" Oct 13 13:19:40 crc kubenswrapper[4678]: I1013 13:19:40.604991 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" path="/var/lib/kubelet/pods/f4be8bfa-2d9f-490b-b239-852793e85b6b/volumes" Oct 13 13:19:55 crc kubenswrapper[4678]: I1013 13:19:55.506663 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:19:55 crc kubenswrapper[4678]: I1013 13:19:55.507313 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:19:55 crc kubenswrapper[4678]: I1013 13:19:55.507363 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:19:55 crc kubenswrapper[4678]: I1013 13:19:55.508181 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:19:55 crc kubenswrapper[4678]: I1013 13:19:55.508250 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" gracePeriod=600 Oct 13 13:19:55 crc kubenswrapper[4678]: E1013 13:19:55.652103 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:19:56 crc kubenswrapper[4678]: I1013 13:19:56.243133 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" exitCode=0 Oct 13 13:19:56 crc kubenswrapper[4678]: I1013 13:19:56.243192 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d"} Oct 13 13:19:56 crc kubenswrapper[4678]: I1013 13:19:56.243243 4678 scope.go:117] "RemoveContainer" containerID="b9ae427c8f50b7ccac564f45566dd3a767fbe13d1222cb301bbf018029bc5415" Oct 13 13:19:56 crc kubenswrapper[4678]: I1013 13:19:56.244376 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:19:56 crc kubenswrapper[4678]: E1013 13:19:56.245241 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.951648 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:06 crc kubenswrapper[4678]: E1013 13:20:06.952541 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="extract-utilities" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.952553 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="extract-utilities" Oct 13 13:20:06 crc kubenswrapper[4678]: E1013 13:20:06.952581 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="registry-server" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.952587 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="registry-server" Oct 13 13:20:06 crc kubenswrapper[4678]: E1013 13:20:06.952614 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="extract-content" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.952620 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="extract-content" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.952803 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4be8bfa-2d9f-490b-b239-852793e85b6b" containerName="registry-server" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.954077 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:06 crc kubenswrapper[4678]: I1013 13:20:06.966253 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.079793 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxqx5\" (UniqueName: \"kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.080176 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.080220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.182315 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxqx5\" (UniqueName: \"kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.182433 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.182492 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.183009 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.183044 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.211926 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxqx5\" (UniqueName: \"kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5\") pod \"certified-operators-6stpx\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.293462 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:07 crc kubenswrapper[4678]: I1013 13:20:07.813175 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:08 crc kubenswrapper[4678]: I1013 13:20:08.368679 4678 generic.go:334] "Generic (PLEG): container finished" podID="264144b4-165d-4920-85e7-2eff14a38fd1" containerID="87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0" exitCode=0 Oct 13 13:20:08 crc kubenswrapper[4678]: I1013 13:20:08.368744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerDied","Data":"87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0"} Oct 13 13:20:08 crc kubenswrapper[4678]: I1013 13:20:08.370117 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerStarted","Data":"9f3606b0a29d62bdd0cde06ca8c4b240e74788309245f52d824abdf6d414b42c"} Oct 13 13:20:08 crc kubenswrapper[4678]: I1013 13:20:08.371375 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:20:08 crc kubenswrapper[4678]: I1013 13:20:08.592600 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:20:08 crc kubenswrapper[4678]: E1013 13:20:08.592997 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:20:09 crc kubenswrapper[4678]: I1013 13:20:09.382575 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerStarted","Data":"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6"} Oct 13 13:20:10 crc kubenswrapper[4678]: I1013 13:20:10.393430 4678 generic.go:334] "Generic (PLEG): container finished" podID="264144b4-165d-4920-85e7-2eff14a38fd1" containerID="eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6" exitCode=0 Oct 13 13:20:10 crc kubenswrapper[4678]: I1013 13:20:10.393586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerDied","Data":"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6"} Oct 13 13:20:12 crc kubenswrapper[4678]: I1013 13:20:12.438009 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerStarted","Data":"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57"} Oct 13 13:20:12 crc kubenswrapper[4678]: I1013 13:20:12.476152 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6stpx" podStartSLOduration=3.600370661 podStartE2EDuration="6.476132242s" podCreationTimestamp="2025-10-13 13:20:06 +0000 UTC" firstStartedPulling="2025-10-13 13:20:08.371074685 +0000 UTC m=+2196.455612569" lastFinishedPulling="2025-10-13 13:20:11.246836256 +0000 UTC m=+2199.331374150" observedRunningTime="2025-10-13 13:20:12.472545949 +0000 UTC m=+2200.557083873" watchObservedRunningTime="2025-10-13 13:20:12.476132242 +0000 UTC m=+2200.560670146" Oct 13 13:20:17 crc kubenswrapper[4678]: I1013 13:20:17.294738 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:17 crc kubenswrapper[4678]: I1013 13:20:17.295337 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:17 crc kubenswrapper[4678]: I1013 13:20:17.349728 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:17 crc kubenswrapper[4678]: I1013 13:20:17.545739 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:17 crc kubenswrapper[4678]: I1013 13:20:17.598281 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:19 crc kubenswrapper[4678]: I1013 13:20:19.506558 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6stpx" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="registry-server" containerID="cri-o://e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57" gracePeriod=2 Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.468794 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.524974 4678 generic.go:334] "Generic (PLEG): container finished" podID="264144b4-165d-4920-85e7-2eff14a38fd1" containerID="e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57" exitCode=0 Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.525020 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerDied","Data":"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57"} Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.525088 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6stpx" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.525092 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6stpx" event={"ID":"264144b4-165d-4920-85e7-2eff14a38fd1","Type":"ContainerDied","Data":"9f3606b0a29d62bdd0cde06ca8c4b240e74788309245f52d824abdf6d414b42c"} Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.525170 4678 scope.go:117] "RemoveContainer" containerID="e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.545626 4678 scope.go:117] "RemoveContainer" containerID="eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.565455 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content\") pod \"264144b4-165d-4920-85e7-2eff14a38fd1\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.565597 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxqx5\" (UniqueName: \"kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5\") pod \"264144b4-165d-4920-85e7-2eff14a38fd1\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.565832 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities\") pod \"264144b4-165d-4920-85e7-2eff14a38fd1\" (UID: \"264144b4-165d-4920-85e7-2eff14a38fd1\") " Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.566806 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities" (OuterVolumeSpecName: "utilities") pod "264144b4-165d-4920-85e7-2eff14a38fd1" (UID: "264144b4-165d-4920-85e7-2eff14a38fd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.567566 4678 scope.go:117] "RemoveContainer" containerID="87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.572167 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5" (OuterVolumeSpecName: "kube-api-access-hxqx5") pod "264144b4-165d-4920-85e7-2eff14a38fd1" (UID: "264144b4-165d-4920-85e7-2eff14a38fd1"). InnerVolumeSpecName "kube-api-access-hxqx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.610275 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "264144b4-165d-4920-85e7-2eff14a38fd1" (UID: "264144b4-165d-4920-85e7-2eff14a38fd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.666310 4678 scope.go:117] "RemoveContainer" containerID="e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57" Oct 13 13:20:20 crc kubenswrapper[4678]: E1013 13:20:20.666742 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57\": container with ID starting with e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57 not found: ID does not exist" containerID="e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.666780 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57"} err="failed to get container status \"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57\": rpc error: code = NotFound desc = could not find container \"e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57\": container with ID starting with e3971e93e02e086df4e5fae94da474416ee0972b58016b703e8a741a777a3f57 not found: ID does not exist" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.666800 4678 scope.go:117] "RemoveContainer" containerID="eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6" Oct 13 13:20:20 crc kubenswrapper[4678]: E1013 13:20:20.667178 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6\": container with ID starting with eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6 not found: ID does not exist" containerID="eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.667234 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6"} err="failed to get container status \"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6\": rpc error: code = NotFound desc = could not find container \"eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6\": container with ID starting with eab566cbe0ef83a406a2434c3c83c80fb74aa3dbf6691f8ef3cb17783fefb3f6 not found: ID does not exist" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.667267 4678 scope.go:117] "RemoveContainer" containerID="87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0" Oct 13 13:20:20 crc kubenswrapper[4678]: E1013 13:20:20.667518 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0\": container with ID starting with 87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0 not found: ID does not exist" containerID="87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.667542 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0"} err="failed to get container status \"87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0\": rpc error: code = NotFound desc = could not find container \"87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0\": container with ID starting with 87dcbeee5b6dd2fa9ec15287f0252808144beb4847a9c620d06df9d2b89d2fb0 not found: ID does not exist" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.668562 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.668602 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264144b4-165d-4920-85e7-2eff14a38fd1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.668622 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxqx5\" (UniqueName: \"kubernetes.io/projected/264144b4-165d-4920-85e7-2eff14a38fd1-kube-api-access-hxqx5\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.857603 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:20 crc kubenswrapper[4678]: I1013 13:20:20.866335 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6stpx"] Oct 13 13:20:22 crc kubenswrapper[4678]: I1013 13:20:22.610427 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" path="/var/lib/kubelet/pods/264144b4-165d-4920-85e7-2eff14a38fd1/volumes" Oct 13 13:20:23 crc kubenswrapper[4678]: I1013 13:20:23.592851 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:20:23 crc kubenswrapper[4678]: E1013 13:20:23.593160 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:20:34 crc kubenswrapper[4678]: I1013 13:20:34.592861 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:20:34 crc kubenswrapper[4678]: E1013 13:20:34.593613 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:20:49 crc kubenswrapper[4678]: I1013 13:20:49.593701 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:20:49 crc kubenswrapper[4678]: E1013 13:20:49.595278 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:20:53 crc kubenswrapper[4678]: I1013 13:20:53.809007 4678 generic.go:334] "Generic (PLEG): container finished" podID="08cf88a5-76f1-455d-a838-9d8e37902c6e" containerID="75d35c54549e9e0d14023656a4fe037798b03719e074ef0739b9baed70f17200" exitCode=0 Oct 13 13:20:53 crc kubenswrapper[4678]: I1013 13:20:53.809376 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" event={"ID":"08cf88a5-76f1-455d-a838-9d8e37902c6e","Type":"ContainerDied","Data":"75d35c54549e9e0d14023656a4fe037798b03719e074ef0739b9baed70f17200"} Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.298947 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.427571 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle\") pod \"08cf88a5-76f1-455d-a838-9d8e37902c6e\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.427881 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0\") pod \"08cf88a5-76f1-455d-a838-9d8e37902c6e\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.428003 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jl29\" (UniqueName: \"kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29\") pod \"08cf88a5-76f1-455d-a838-9d8e37902c6e\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.428153 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key\") pod \"08cf88a5-76f1-455d-a838-9d8e37902c6e\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.428270 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory\") pod \"08cf88a5-76f1-455d-a838-9d8e37902c6e\" (UID: \"08cf88a5-76f1-455d-a838-9d8e37902c6e\") " Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.432651 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "08cf88a5-76f1-455d-a838-9d8e37902c6e" (UID: "08cf88a5-76f1-455d-a838-9d8e37902c6e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.435973 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29" (OuterVolumeSpecName: "kube-api-access-8jl29") pod "08cf88a5-76f1-455d-a838-9d8e37902c6e" (UID: "08cf88a5-76f1-455d-a838-9d8e37902c6e"). InnerVolumeSpecName "kube-api-access-8jl29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.456895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "08cf88a5-76f1-455d-a838-9d8e37902c6e" (UID: "08cf88a5-76f1-455d-a838-9d8e37902c6e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.464862 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory" (OuterVolumeSpecName: "inventory") pod "08cf88a5-76f1-455d-a838-9d8e37902c6e" (UID: "08cf88a5-76f1-455d-a838-9d8e37902c6e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.465631 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "08cf88a5-76f1-455d-a838-9d8e37902c6e" (UID: "08cf88a5-76f1-455d-a838-9d8e37902c6e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.530624 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.530664 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.530675 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.530687 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/08cf88a5-76f1-455d-a838-9d8e37902c6e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.530696 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jl29\" (UniqueName: \"kubernetes.io/projected/08cf88a5-76f1-455d-a838-9d8e37902c6e-kube-api-access-8jl29\") on node \"crc\" DevicePath \"\"" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.827185 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" event={"ID":"08cf88a5-76f1-455d-a838-9d8e37902c6e","Type":"ContainerDied","Data":"cf1c55340ab248f4d8f0181927251ef82b4bd6380e557e38cc9eec7d178a07eb"} Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.827231 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf1c55340ab248f4d8f0181927251ef82b4bd6380e557e38cc9eec7d178a07eb" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.827340 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-m62j6" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.943636 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm"] Oct 13 13:20:55 crc kubenswrapper[4678]: E1013 13:20:55.944275 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="extract-content" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944303 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="extract-content" Oct 13 13:20:55 crc kubenswrapper[4678]: E1013 13:20:55.944347 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08cf88a5-76f1-455d-a838-9d8e37902c6e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944360 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="08cf88a5-76f1-455d-a838-9d8e37902c6e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 13:20:55 crc kubenswrapper[4678]: E1013 13:20:55.944401 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="extract-utilities" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944412 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="extract-utilities" Oct 13 13:20:55 crc kubenswrapper[4678]: E1013 13:20:55.944430 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="registry-server" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944440 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="registry-server" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944785 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="08cf88a5-76f1-455d-a838-9d8e37902c6e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.944818 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="264144b4-165d-4920-85e7-2eff14a38fd1" containerName="registry-server" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.947362 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.952926 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.952942 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.952948 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.953259 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.953269 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.953280 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.953280 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 13 13:20:55 crc kubenswrapper[4678]: I1013 13:20:55.961680 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm"] Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.142648 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.142934 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.142978 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.142996 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.143024 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.143071 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.143101 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.143140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.143405 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dfkp\" (UniqueName: \"kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245524 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245551 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245620 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dfkp\" (UniqueName: \"kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245685 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245726 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.245785 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.247116 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.250970 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.251699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.251709 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.251703 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.253119 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.261172 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.261539 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.264307 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dfkp\" (UniqueName: \"kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w28hm\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.268230 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.804542 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm"] Oct 13 13:20:56 crc kubenswrapper[4678]: I1013 13:20:56.841335 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" event={"ID":"204374ee-a5e4-4a02-bada-f57883e5c571","Type":"ContainerStarted","Data":"6b4663a01bb778102e82c4bd8acf7927b9d47b7d0db2069cbfd9e631ab1c17bf"} Oct 13 13:20:57 crc kubenswrapper[4678]: I1013 13:20:57.852476 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" event={"ID":"204374ee-a5e4-4a02-bada-f57883e5c571","Type":"ContainerStarted","Data":"5886843de3ef23b0c4107625725032ecb2c63316200b4110540131640ed92282"} Oct 13 13:20:57 crc kubenswrapper[4678]: I1013 13:20:57.871259 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" podStartSLOduration=2.248699405 podStartE2EDuration="2.871244098s" podCreationTimestamp="2025-10-13 13:20:55 +0000 UTC" firstStartedPulling="2025-10-13 13:20:56.813154333 +0000 UTC m=+2244.897692217" lastFinishedPulling="2025-10-13 13:20:57.435699026 +0000 UTC m=+2245.520236910" observedRunningTime="2025-10-13 13:20:57.866493851 +0000 UTC m=+2245.951031735" watchObservedRunningTime="2025-10-13 13:20:57.871244098 +0000 UTC m=+2245.955781982" Oct 13 13:21:03 crc kubenswrapper[4678]: I1013 13:21:03.592137 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:21:03 crc kubenswrapper[4678]: E1013 13:21:03.594419 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:21:14 crc kubenswrapper[4678]: I1013 13:21:14.593262 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:21:14 crc kubenswrapper[4678]: E1013 13:21:14.593976 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:21:28 crc kubenswrapper[4678]: I1013 13:21:28.592531 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:21:28 crc kubenswrapper[4678]: E1013 13:21:28.593274 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:21:43 crc kubenswrapper[4678]: I1013 13:21:43.593389 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:21:43 crc kubenswrapper[4678]: E1013 13:21:43.594524 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:21:54 crc kubenswrapper[4678]: I1013 13:21:54.592984 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:21:54 crc kubenswrapper[4678]: E1013 13:21:54.593952 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:22:09 crc kubenswrapper[4678]: I1013 13:22:09.592317 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:22:09 crc kubenswrapper[4678]: E1013 13:22:09.593147 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:22:24 crc kubenswrapper[4678]: I1013 13:22:24.594617 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:22:24 crc kubenswrapper[4678]: E1013 13:22:24.595995 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:22:35 crc kubenswrapper[4678]: I1013 13:22:35.592948 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:22:35 crc kubenswrapper[4678]: E1013 13:22:35.594037 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:22:50 crc kubenswrapper[4678]: I1013 13:22:50.592365 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:22:50 crc kubenswrapper[4678]: E1013 13:22:50.592899 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:23:04 crc kubenswrapper[4678]: I1013 13:23:04.592935 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:23:04 crc kubenswrapper[4678]: E1013 13:23:04.595054 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:23:17 crc kubenswrapper[4678]: I1013 13:23:17.592943 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:23:17 crc kubenswrapper[4678]: E1013 13:23:17.593826 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:23:29 crc kubenswrapper[4678]: I1013 13:23:29.592788 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:23:29 crc kubenswrapper[4678]: E1013 13:23:29.593549 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:23:39 crc kubenswrapper[4678]: I1013 13:23:39.853867 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:39 crc kubenswrapper[4678]: I1013 13:23:39.856703 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:39 crc kubenswrapper[4678]: I1013 13:23:39.881704 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.007861 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.008989 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vq7f\" (UniqueName: \"kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.009190 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.111229 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.111334 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.111440 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vq7f\" (UniqueName: \"kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.112527 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.112561 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.131232 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vq7f\" (UniqueName: \"kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f\") pod \"redhat-marketplace-27fk8\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.196331 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:40 crc kubenswrapper[4678]: I1013 13:23:40.719552 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:41 crc kubenswrapper[4678]: I1013 13:23:41.361658 4678 generic.go:334] "Generic (PLEG): container finished" podID="92fddc45-3b33-49c8-9e14-f83712c42240" containerID="ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0" exitCode=0 Oct 13 13:23:41 crc kubenswrapper[4678]: I1013 13:23:41.361707 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerDied","Data":"ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0"} Oct 13 13:23:41 crc kubenswrapper[4678]: I1013 13:23:41.362094 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerStarted","Data":"41dab482c7b98a3f15d2246f02405972cc7f67ca74c7137698e24b37846b8931"} Oct 13 13:23:43 crc kubenswrapper[4678]: I1013 13:23:43.382757 4678 generic.go:334] "Generic (PLEG): container finished" podID="92fddc45-3b33-49c8-9e14-f83712c42240" containerID="7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25" exitCode=0 Oct 13 13:23:43 crc kubenswrapper[4678]: I1013 13:23:43.382850 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerDied","Data":"7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25"} Oct 13 13:23:43 crc kubenswrapper[4678]: I1013 13:23:43.592236 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:23:43 crc kubenswrapper[4678]: E1013 13:23:43.592616 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:23:44 crc kubenswrapper[4678]: I1013 13:23:44.393691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerStarted","Data":"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4"} Oct 13 13:23:44 crc kubenswrapper[4678]: I1013 13:23:44.418527 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-27fk8" podStartSLOduration=2.667588779 podStartE2EDuration="5.418495606s" podCreationTimestamp="2025-10-13 13:23:39 +0000 UTC" firstStartedPulling="2025-10-13 13:23:41.364098385 +0000 UTC m=+2409.448636269" lastFinishedPulling="2025-10-13 13:23:44.115005212 +0000 UTC m=+2412.199543096" observedRunningTime="2025-10-13 13:23:44.411278644 +0000 UTC m=+2412.495816528" watchObservedRunningTime="2025-10-13 13:23:44.418495606 +0000 UTC m=+2412.503033510" Oct 13 13:23:50 crc kubenswrapper[4678]: I1013 13:23:50.196937 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:50 crc kubenswrapper[4678]: I1013 13:23:50.197517 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:50 crc kubenswrapper[4678]: I1013 13:23:50.268516 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:50 crc kubenswrapper[4678]: I1013 13:23:50.557107 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:50 crc kubenswrapper[4678]: I1013 13:23:50.612896 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.483457 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-27fk8" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="registry-server" containerID="cri-o://f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4" gracePeriod=2 Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.963449 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.979205 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vq7f\" (UniqueName: \"kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f\") pod \"92fddc45-3b33-49c8-9e14-f83712c42240\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.979275 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content\") pod \"92fddc45-3b33-49c8-9e14-f83712c42240\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.979406 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities\") pod \"92fddc45-3b33-49c8-9e14-f83712c42240\" (UID: \"92fddc45-3b33-49c8-9e14-f83712c42240\") " Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.981852 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities" (OuterVolumeSpecName: "utilities") pod "92fddc45-3b33-49c8-9e14-f83712c42240" (UID: "92fddc45-3b33-49c8-9e14-f83712c42240"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.987376 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f" (OuterVolumeSpecName: "kube-api-access-8vq7f") pod "92fddc45-3b33-49c8-9e14-f83712c42240" (UID: "92fddc45-3b33-49c8-9e14-f83712c42240"). InnerVolumeSpecName "kube-api-access-8vq7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:23:52 crc kubenswrapper[4678]: I1013 13:23:52.996802 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92fddc45-3b33-49c8-9e14-f83712c42240" (UID: "92fddc45-3b33-49c8-9e14-f83712c42240"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.080818 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.080851 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vq7f\" (UniqueName: \"kubernetes.io/projected/92fddc45-3b33-49c8-9e14-f83712c42240-kube-api-access-8vq7f\") on node \"crc\" DevicePath \"\"" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.080861 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92fddc45-3b33-49c8-9e14-f83712c42240-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.495636 4678 generic.go:334] "Generic (PLEG): container finished" podID="92fddc45-3b33-49c8-9e14-f83712c42240" containerID="f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4" exitCode=0 Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.495701 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerDied","Data":"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4"} Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.495714 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-27fk8" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.495747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-27fk8" event={"ID":"92fddc45-3b33-49c8-9e14-f83712c42240","Type":"ContainerDied","Data":"41dab482c7b98a3f15d2246f02405972cc7f67ca74c7137698e24b37846b8931"} Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.495766 4678 scope.go:117] "RemoveContainer" containerID="f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.528441 4678 scope.go:117] "RemoveContainer" containerID="7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.543127 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.551479 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-27fk8"] Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.557497 4678 scope.go:117] "RemoveContainer" containerID="ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.598343 4678 scope.go:117] "RemoveContainer" containerID="f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4" Oct 13 13:23:53 crc kubenswrapper[4678]: E1013 13:23:53.598783 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4\": container with ID starting with f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4 not found: ID does not exist" containerID="f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.598853 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4"} err="failed to get container status \"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4\": rpc error: code = NotFound desc = could not find container \"f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4\": container with ID starting with f04bde6648140d8413211f9017a055a9d8383641c54235bfea4dc6df91640de4 not found: ID does not exist" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.598892 4678 scope.go:117] "RemoveContainer" containerID="7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25" Oct 13 13:23:53 crc kubenswrapper[4678]: E1013 13:23:53.599239 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25\": container with ID starting with 7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25 not found: ID does not exist" containerID="7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.599286 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25"} err="failed to get container status \"7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25\": rpc error: code = NotFound desc = could not find container \"7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25\": container with ID starting with 7e6f6fd7a7987a0b343cd2c4ca8d3961ca72adba3b80dae8982f04c2e76b7c25 not found: ID does not exist" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.599312 4678 scope.go:117] "RemoveContainer" containerID="ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0" Oct 13 13:23:53 crc kubenswrapper[4678]: E1013 13:23:53.599571 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0\": container with ID starting with ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0 not found: ID does not exist" containerID="ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0" Oct 13 13:23:53 crc kubenswrapper[4678]: I1013 13:23:53.599709 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0"} err="failed to get container status \"ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0\": rpc error: code = NotFound desc = could not find container \"ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0\": container with ID starting with ec0e393e0ea1009caddfb4ae1fa4910fb90d7fa75e8dde219c1d1090d15c4ac0 not found: ID does not exist" Oct 13 13:23:54 crc kubenswrapper[4678]: I1013 13:23:54.603461 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" path="/var/lib/kubelet/pods/92fddc45-3b33-49c8-9e14-f83712c42240/volumes" Oct 13 13:23:57 crc kubenswrapper[4678]: I1013 13:23:57.593183 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:23:57 crc kubenswrapper[4678]: E1013 13:23:57.594160 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:24:06 crc kubenswrapper[4678]: I1013 13:24:06.648165 4678 generic.go:334] "Generic (PLEG): container finished" podID="204374ee-a5e4-4a02-bada-f57883e5c571" containerID="5886843de3ef23b0c4107625725032ecb2c63316200b4110540131640ed92282" exitCode=0 Oct 13 13:24:06 crc kubenswrapper[4678]: I1013 13:24:06.648272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" event={"ID":"204374ee-a5e4-4a02-bada-f57883e5c571","Type":"ContainerDied","Data":"5886843de3ef23b0c4107625725032ecb2c63316200b4110540131640ed92282"} Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.201278 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.289975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290725 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290761 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290792 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290814 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dfkp\" (UniqueName: \"kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290887 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290938 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.290999 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory\") pod \"204374ee-a5e4-4a02-bada-f57883e5c571\" (UID: \"204374ee-a5e4-4a02-bada-f57883e5c571\") " Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.295616 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.301214 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp" (OuterVolumeSpecName: "kube-api-access-7dfkp") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "kube-api-access-7dfkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.313688 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.321689 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory" (OuterVolumeSpecName: "inventory") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.322322 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.322349 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.322892 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.324360 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.325224 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "204374ee-a5e4-4a02-bada-f57883e5c571" (UID: "204374ee-a5e4-4a02-bada-f57883e5c571"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393093 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dfkp\" (UniqueName: \"kubernetes.io/projected/204374ee-a5e4-4a02-bada-f57883e5c571-kube-api-access-7dfkp\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393131 4678 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393144 4678 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393156 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393167 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393182 4678 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393192 4678 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393204 4678 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/204374ee-a5e4-4a02-bada-f57883e5c571-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.393215 4678 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/204374ee-a5e4-4a02-bada-f57883e5c571-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.669434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" event={"ID":"204374ee-a5e4-4a02-bada-f57883e5c571","Type":"ContainerDied","Data":"6b4663a01bb778102e82c4bd8acf7927b9d47b7d0db2069cbfd9e631ab1c17bf"} Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.669487 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b4663a01bb778102e82c4bd8acf7927b9d47b7d0db2069cbfd9e631ab1c17bf" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.669505 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w28hm" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769040 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2"] Oct 13 13:24:08 crc kubenswrapper[4678]: E1013 13:24:08.769620 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="registry-server" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769646 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="registry-server" Oct 13 13:24:08 crc kubenswrapper[4678]: E1013 13:24:08.769687 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="extract-utilities" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769697 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="extract-utilities" Oct 13 13:24:08 crc kubenswrapper[4678]: E1013 13:24:08.769718 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="204374ee-a5e4-4a02-bada-f57883e5c571" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769727 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="204374ee-a5e4-4a02-bada-f57883e5c571" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 13:24:08 crc kubenswrapper[4678]: E1013 13:24:08.769743 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="extract-content" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769751 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="extract-content" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.769971 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="204374ee-a5e4-4a02-bada-f57883e5c571" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.770004 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="92fddc45-3b33-49c8-9e14-f83712c42240" containerName="registry-server" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.770649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.773082 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.773096 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.773111 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.774435 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.774447 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-9vj98" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.779545 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2"] Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903549 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903755 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m4l9\" (UniqueName: \"kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903891 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:08 crc kubenswrapper[4678]: I1013 13:24:08.903974 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.006557 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.006678 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.006723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.006787 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.007743 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.007911 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m4l9\" (UniqueName: \"kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.007957 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.011366 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.011406 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.012405 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.012875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.014284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.014308 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.028844 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m4l9\" (UniqueName: \"kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-x56k2\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.135702 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:24:09 crc kubenswrapper[4678]: I1013 13:24:09.666594 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2"] Oct 13 13:24:10 crc kubenswrapper[4678]: I1013 13:24:10.592151 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:24:10 crc kubenswrapper[4678]: E1013 13:24:10.592935 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:24:10 crc kubenswrapper[4678]: I1013 13:24:10.709208 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" event={"ID":"d7940e67-463a-4cc5-812e-0bde560d6eda","Type":"ContainerStarted","Data":"22492af76cc661f717942459f787b493483866babdc44a7fedfecd5c42003fc6"} Oct 13 13:24:10 crc kubenswrapper[4678]: I1013 13:24:10.709280 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" event={"ID":"d7940e67-463a-4cc5-812e-0bde560d6eda","Type":"ContainerStarted","Data":"c71ff580c3d030648fb9ea13e70782920a38f3854bb035d5a5f69c0ab67378f7"} Oct 13 13:24:10 crc kubenswrapper[4678]: I1013 13:24:10.744341 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" podStartSLOduration=2.185474725 podStartE2EDuration="2.744307551s" podCreationTimestamp="2025-10-13 13:24:08 +0000 UTC" firstStartedPulling="2025-10-13 13:24:09.686398811 +0000 UTC m=+2437.770936695" lastFinishedPulling="2025-10-13 13:24:10.245231607 +0000 UTC m=+2438.329769521" observedRunningTime="2025-10-13 13:24:10.733836392 +0000 UTC m=+2438.818374296" watchObservedRunningTime="2025-10-13 13:24:10.744307551 +0000 UTC m=+2438.828845485" Oct 13 13:24:21 crc kubenswrapper[4678]: I1013 13:24:21.592888 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:24:21 crc kubenswrapper[4678]: E1013 13:24:21.593685 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:24:32 crc kubenswrapper[4678]: I1013 13:24:32.598252 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:24:32 crc kubenswrapper[4678]: E1013 13:24:32.599134 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:24:43 crc kubenswrapper[4678]: I1013 13:24:43.593148 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:24:43 crc kubenswrapper[4678]: E1013 13:24:43.594509 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:24:54 crc kubenswrapper[4678]: I1013 13:24:54.593911 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:24:54 crc kubenswrapper[4678]: E1013 13:24:54.595554 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:25:05 crc kubenswrapper[4678]: I1013 13:25:05.593394 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:25:06 crc kubenswrapper[4678]: I1013 13:25:06.220868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b"} Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.199216 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.202737 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.215291 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.255297 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5bnr\" (UniqueName: \"kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.255433 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.255458 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.357633 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5bnr\" (UniqueName: \"kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.357713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.357735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.358266 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.358378 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.380866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5bnr\" (UniqueName: \"kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr\") pod \"redhat-operators-t95kj\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:26 crc kubenswrapper[4678]: I1013 13:26:26.538179 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:27 crc kubenswrapper[4678]: I1013 13:26:27.092407 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:27 crc kubenswrapper[4678]: I1013 13:26:27.941981 4678 generic.go:334] "Generic (PLEG): container finished" podID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerID="dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259" exitCode=0 Oct 13 13:26:27 crc kubenswrapper[4678]: I1013 13:26:27.942107 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerDied","Data":"dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259"} Oct 13 13:26:27 crc kubenswrapper[4678]: I1013 13:26:27.942267 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerStarted","Data":"dc611e211826ed1b6d95b2990369fb0c4650f75d8898aa15723c013fb86aff0a"} Oct 13 13:26:27 crc kubenswrapper[4678]: I1013 13:26:27.948364 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:26:29 crc kubenswrapper[4678]: I1013 13:26:29.961504 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerStarted","Data":"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60"} Oct 13 13:26:30 crc kubenswrapper[4678]: I1013 13:26:30.976655 4678 generic.go:334] "Generic (PLEG): container finished" podID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerID="65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60" exitCode=0 Oct 13 13:26:30 crc kubenswrapper[4678]: I1013 13:26:30.976885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerDied","Data":"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60"} Oct 13 13:26:33 crc kubenswrapper[4678]: I1013 13:26:33.001565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerStarted","Data":"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0"} Oct 13 13:26:33 crc kubenswrapper[4678]: I1013 13:26:33.025946 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t95kj" podStartSLOduration=3.193705541 podStartE2EDuration="7.025928678s" podCreationTimestamp="2025-10-13 13:26:26 +0000 UTC" firstStartedPulling="2025-10-13 13:26:27.947672956 +0000 UTC m=+2576.032210880" lastFinishedPulling="2025-10-13 13:26:31.779896123 +0000 UTC m=+2579.864434017" observedRunningTime="2025-10-13 13:26:33.019740727 +0000 UTC m=+2581.104278611" watchObservedRunningTime="2025-10-13 13:26:33.025928678 +0000 UTC m=+2581.110466562" Oct 13 13:26:35 crc kubenswrapper[4678]: I1013 13:26:35.017176 4678 generic.go:334] "Generic (PLEG): container finished" podID="d7940e67-463a-4cc5-812e-0bde560d6eda" containerID="22492af76cc661f717942459f787b493483866babdc44a7fedfecd5c42003fc6" exitCode=0 Oct 13 13:26:35 crc kubenswrapper[4678]: I1013 13:26:35.017244 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" event={"ID":"d7940e67-463a-4cc5-812e-0bde560d6eda","Type":"ContainerDied","Data":"22492af76cc661f717942459f787b493483866babdc44a7fedfecd5c42003fc6"} Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.451237 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.538832 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.538894 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.568650 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.568872 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.568939 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.568975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.569015 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m4l9\" (UniqueName: \"kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.569051 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.569109 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2\") pod \"d7940e67-463a-4cc5-812e-0bde560d6eda\" (UID: \"d7940e67-463a-4cc5-812e-0bde560d6eda\") " Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.583990 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9" (OuterVolumeSpecName: "kube-api-access-8m4l9") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "kube-api-access-8m4l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.584237 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.584657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.602820 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.605957 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.612093 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.613341 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.614117 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory" (OuterVolumeSpecName: "inventory") pod "d7940e67-463a-4cc5-812e-0bde560d6eda" (UID: "d7940e67-463a-4cc5-812e-0bde560d6eda"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671898 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671936 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-inventory\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671949 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671960 4678 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671970 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m4l9\" (UniqueName: \"kubernetes.io/projected/d7940e67-463a-4cc5-812e-0bde560d6eda-kube-api-access-8m4l9\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671979 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:36 crc kubenswrapper[4678]: I1013 13:26:36.671987 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d7940e67-463a-4cc5-812e-0bde560d6eda-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:37 crc kubenswrapper[4678]: I1013 13:26:37.051774 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" Oct 13 13:26:37 crc kubenswrapper[4678]: I1013 13:26:37.051955 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-x56k2" event={"ID":"d7940e67-463a-4cc5-812e-0bde560d6eda","Type":"ContainerDied","Data":"c71ff580c3d030648fb9ea13e70782920a38f3854bb035d5a5f69c0ab67378f7"} Oct 13 13:26:37 crc kubenswrapper[4678]: I1013 13:26:37.052483 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c71ff580c3d030648fb9ea13e70782920a38f3854bb035d5a5f69c0ab67378f7" Oct 13 13:26:37 crc kubenswrapper[4678]: I1013 13:26:37.111860 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:37 crc kubenswrapper[4678]: E1013 13:26:37.149621 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7940e67_463a_4cc5_812e_0bde560d6eda.slice\": RecentStats: unable to find data in memory cache]" Oct 13 13:26:37 crc kubenswrapper[4678]: I1013 13:26:37.167997 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.065916 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t95kj" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="registry-server" containerID="cri-o://972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0" gracePeriod=2 Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.547127 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.634013 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content\") pod \"0661670f-da74-4ef4-ae6d-435b330c00d3\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.634123 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities\") pod \"0661670f-da74-4ef4-ae6d-435b330c00d3\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.634244 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5bnr\" (UniqueName: \"kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr\") pod \"0661670f-da74-4ef4-ae6d-435b330c00d3\" (UID: \"0661670f-da74-4ef4-ae6d-435b330c00d3\") " Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.635493 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities" (OuterVolumeSpecName: "utilities") pod "0661670f-da74-4ef4-ae6d-435b330c00d3" (UID: "0661670f-da74-4ef4-ae6d-435b330c00d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.642333 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr" (OuterVolumeSpecName: "kube-api-access-l5bnr") pod "0661670f-da74-4ef4-ae6d-435b330c00d3" (UID: "0661670f-da74-4ef4-ae6d-435b330c00d3"). InnerVolumeSpecName "kube-api-access-l5bnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.648030 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.731993 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0661670f-da74-4ef4-ae6d-435b330c00d3" (UID: "0661670f-da74-4ef4-ae6d-435b330c00d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.750509 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661670f-da74-4ef4-ae6d-435b330c00d3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:39 crc kubenswrapper[4678]: I1013 13:26:39.750561 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5bnr\" (UniqueName: \"kubernetes.io/projected/0661670f-da74-4ef4-ae6d-435b330c00d3-kube-api-access-l5bnr\") on node \"crc\" DevicePath \"\"" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.081138 4678 generic.go:334] "Generic (PLEG): container finished" podID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerID="972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0" exitCode=0 Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.081194 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerDied","Data":"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0"} Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.081230 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t95kj" event={"ID":"0661670f-da74-4ef4-ae6d-435b330c00d3","Type":"ContainerDied","Data":"dc611e211826ed1b6d95b2990369fb0c4650f75d8898aa15723c013fb86aff0a"} Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.081247 4678 scope.go:117] "RemoveContainer" containerID="972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.082112 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t95kj" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.104979 4678 scope.go:117] "RemoveContainer" containerID="65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.117162 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.125743 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t95kj"] Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.154397 4678 scope.go:117] "RemoveContainer" containerID="dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.187494 4678 scope.go:117] "RemoveContainer" containerID="972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0" Oct 13 13:26:40 crc kubenswrapper[4678]: E1013 13:26:40.188130 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0\": container with ID starting with 972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0 not found: ID does not exist" containerID="972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.188200 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0"} err="failed to get container status \"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0\": rpc error: code = NotFound desc = could not find container \"972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0\": container with ID starting with 972f1e2053df80e64b64fc9b69fe4696923b2c02e0e5fed96de607996af7c8f0 not found: ID does not exist" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.188247 4678 scope.go:117] "RemoveContainer" containerID="65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60" Oct 13 13:26:40 crc kubenswrapper[4678]: E1013 13:26:40.188933 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60\": container with ID starting with 65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60 not found: ID does not exist" containerID="65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.188988 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60"} err="failed to get container status \"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60\": rpc error: code = NotFound desc = could not find container \"65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60\": container with ID starting with 65c2c75d9489cbc05d1f6c0c19af0e18b29f738794d1915507585cdc9a57ca60 not found: ID does not exist" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.189018 4678 scope.go:117] "RemoveContainer" containerID="dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259" Oct 13 13:26:40 crc kubenswrapper[4678]: E1013 13:26:40.189596 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259\": container with ID starting with dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259 not found: ID does not exist" containerID="dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.189637 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259"} err="failed to get container status \"dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259\": rpc error: code = NotFound desc = could not find container \"dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259\": container with ID starting with dc18d8bacb0ba40d31ea9fb41e60414bba7fd38a692cf00a040bb7db7a996259 not found: ID does not exist" Oct 13 13:26:40 crc kubenswrapper[4678]: I1013 13:26:40.607143 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" path="/var/lib/kubelet/pods/0661670f-da74-4ef4-ae6d-435b330c00d3/volumes" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.987179 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 13:27:21 crc kubenswrapper[4678]: E1013 13:27:21.988261 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7940e67-463a-4cc5-812e-0bde560d6eda" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988277 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7940e67-463a-4cc5-812e-0bde560d6eda" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 13:27:21 crc kubenswrapper[4678]: E1013 13:27:21.988292 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="extract-content" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988298 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="extract-content" Oct 13 13:27:21 crc kubenswrapper[4678]: E1013 13:27:21.988318 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="registry-server" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988324 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="registry-server" Oct 13 13:27:21 crc kubenswrapper[4678]: E1013 13:27:21.988355 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="extract-utilities" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988363 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="extract-utilities" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988532 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0661670f-da74-4ef4-ae6d-435b330c00d3" containerName="registry-server" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.988547 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7940e67-463a-4cc5-812e-0bde560d6eda" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.989174 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.991740 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.991874 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-z78sb" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.993523 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 13 13:27:21 crc kubenswrapper[4678]: I1013 13:27:21.994738 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.007685 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.050847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051056 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051169 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051223 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051275 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051357 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051475 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051535 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pdgn\" (UniqueName: \"kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.051593 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153314 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pdgn\" (UniqueName: \"kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153410 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153493 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153570 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153627 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153656 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153700 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.153762 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.154006 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.154124 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.154213 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.155010 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.155348 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.159812 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.160593 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.166713 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.169221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pdgn\" (UniqueName: \"kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.193512 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.308964 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 13:27:22 crc kubenswrapper[4678]: I1013 13:27:22.730257 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 13 13:27:23 crc kubenswrapper[4678]: I1013 13:27:23.486655 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a","Type":"ContainerStarted","Data":"98f8052521704e97feb99d69d89d649d40e40a90c70bdbc5210458c3d700b277"} Oct 13 13:27:25 crc kubenswrapper[4678]: I1013 13:27:25.505778 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:27:25 crc kubenswrapper[4678]: I1013 13:27:25.506151 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:27:48 crc kubenswrapper[4678]: E1013 13:27:48.714219 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 13 13:27:48 crc kubenswrapper[4678]: E1013 13:27:48.715182 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5pdgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 13 13:27:48 crc kubenswrapper[4678]: E1013 13:27:48.716647 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" Oct 13 13:27:49 crc kubenswrapper[4678]: E1013 13:27:49.720180 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" Oct 13 13:27:55 crc kubenswrapper[4678]: I1013 13:27:55.506227 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:27:55 crc kubenswrapper[4678]: I1013 13:27:55.507247 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:28:03 crc kubenswrapper[4678]: I1013 13:28:03.133618 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 13 13:28:04 crc kubenswrapper[4678]: I1013 13:28:04.840526 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a","Type":"ContainerStarted","Data":"69cb42d4e30b310352bfe97fda9019951e7460f234adddfc3b0f056de7397b57"} Oct 13 13:28:04 crc kubenswrapper[4678]: I1013 13:28:04.863383 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.465694049 podStartE2EDuration="44.863335656s" podCreationTimestamp="2025-10-13 13:27:20 +0000 UTC" firstStartedPulling="2025-10-13 13:27:22.733427448 +0000 UTC m=+2630.817965332" lastFinishedPulling="2025-10-13 13:28:03.131069035 +0000 UTC m=+2671.215606939" observedRunningTime="2025-10-13 13:28:04.861833157 +0000 UTC m=+2672.946371041" watchObservedRunningTime="2025-10-13 13:28:04.863335656 +0000 UTC m=+2672.947873540" Oct 13 13:28:25 crc kubenswrapper[4678]: I1013 13:28:25.505981 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:28:25 crc kubenswrapper[4678]: I1013 13:28:25.506651 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:28:25 crc kubenswrapper[4678]: I1013 13:28:25.506706 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:28:25 crc kubenswrapper[4678]: I1013 13:28:25.507773 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:28:25 crc kubenswrapper[4678]: I1013 13:28:25.507856 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b" gracePeriod=600 Oct 13 13:28:26 crc kubenswrapper[4678]: I1013 13:28:26.044645 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b" exitCode=0 Oct 13 13:28:26 crc kubenswrapper[4678]: I1013 13:28:26.044836 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b"} Oct 13 13:28:26 crc kubenswrapper[4678]: I1013 13:28:26.045006 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12"} Oct 13 13:28:26 crc kubenswrapper[4678]: I1013 13:28:26.045028 4678 scope.go:117] "RemoveContainer" containerID="ebcd08f0e48b66c654da34ae2931f643897d0745c1ca35e6b131dea2793edc4d" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.148347 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn"] Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.150787 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.154876 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.155133 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.161237 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn"] Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.254092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.254170 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk8mt\" (UniqueName: \"kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.254660 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.356436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.356516 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk8mt\" (UniqueName: \"kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.356666 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.358895 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.371417 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.374091 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk8mt\" (UniqueName: \"kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt\") pod \"collect-profiles-29339370-h9hnn\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.475481 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:00 crc kubenswrapper[4678]: I1013 13:30:00.910254 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn"] Oct 13 13:30:01 crc kubenswrapper[4678]: I1013 13:30:01.921756 4678 generic.go:334] "Generic (PLEG): container finished" podID="e33431a9-54ae-414d-bc80-19bf0170e610" containerID="91b882784fdf106b77c5949e95f6cb41e69d0a2aeb374d3287064ad613c3b1f1" exitCode=0 Oct 13 13:30:01 crc kubenswrapper[4678]: I1013 13:30:01.921815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" event={"ID":"e33431a9-54ae-414d-bc80-19bf0170e610","Type":"ContainerDied","Data":"91b882784fdf106b77c5949e95f6cb41e69d0a2aeb374d3287064ad613c3b1f1"} Oct 13 13:30:01 crc kubenswrapper[4678]: I1013 13:30:01.922083 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" event={"ID":"e33431a9-54ae-414d-bc80-19bf0170e610","Type":"ContainerStarted","Data":"a063622e900a5175858cc47651512ea0269fb98fed86d5e6b84dddd4d6f81a1a"} Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.270522 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.414971 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume\") pod \"e33431a9-54ae-414d-bc80-19bf0170e610\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.415030 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk8mt\" (UniqueName: \"kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt\") pod \"e33431a9-54ae-414d-bc80-19bf0170e610\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.415202 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume\") pod \"e33431a9-54ae-414d-bc80-19bf0170e610\" (UID: \"e33431a9-54ae-414d-bc80-19bf0170e610\") " Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.416141 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume" (OuterVolumeSpecName: "config-volume") pod "e33431a9-54ae-414d-bc80-19bf0170e610" (UID: "e33431a9-54ae-414d-bc80-19bf0170e610"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.421385 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e33431a9-54ae-414d-bc80-19bf0170e610" (UID: "e33431a9-54ae-414d-bc80-19bf0170e610"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.421579 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt" (OuterVolumeSpecName: "kube-api-access-gk8mt") pod "e33431a9-54ae-414d-bc80-19bf0170e610" (UID: "e33431a9-54ae-414d-bc80-19bf0170e610"). InnerVolumeSpecName "kube-api-access-gk8mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.517319 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e33431a9-54ae-414d-bc80-19bf0170e610-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.517346 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk8mt\" (UniqueName: \"kubernetes.io/projected/e33431a9-54ae-414d-bc80-19bf0170e610-kube-api-access-gk8mt\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.517357 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e33431a9-54ae-414d-bc80-19bf0170e610-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.941688 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" event={"ID":"e33431a9-54ae-414d-bc80-19bf0170e610","Type":"ContainerDied","Data":"a063622e900a5175858cc47651512ea0269fb98fed86d5e6b84dddd4d6f81a1a"} Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.942070 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a063622e900a5175858cc47651512ea0269fb98fed86d5e6b84dddd4d6f81a1a" Oct 13 13:30:03 crc kubenswrapper[4678]: I1013 13:30:03.942150 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339370-h9hnn" Oct 13 13:30:04 crc kubenswrapper[4678]: I1013 13:30:04.349005 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc"] Oct 13 13:30:04 crc kubenswrapper[4678]: I1013 13:30:04.356439 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339325-jl5mc"] Oct 13 13:30:04 crc kubenswrapper[4678]: I1013 13:30:04.611590 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="839c0ab6-3764-4934-9d39-a09227958854" path="/var/lib/kubelet/pods/839c0ab6-3764-4934-9d39-a09227958854/volumes" Oct 13 13:30:25 crc kubenswrapper[4678]: I1013 13:30:25.506580 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:30:25 crc kubenswrapper[4678]: I1013 13:30:25.507206 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.136007 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:37 crc kubenswrapper[4678]: E1013 13:30:37.137118 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e33431a9-54ae-414d-bc80-19bf0170e610" containerName="collect-profiles" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.137139 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e33431a9-54ae-414d-bc80-19bf0170e610" containerName="collect-profiles" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.137421 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e33431a9-54ae-414d-bc80-19bf0170e610" containerName="collect-profiles" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.139201 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.148868 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.215670 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.215889 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q994q\" (UniqueName: \"kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.216267 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.317737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.317795 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.317852 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q994q\" (UniqueName: \"kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.318316 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.318370 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.331735 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.333734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.347203 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.350864 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q994q\" (UniqueName: \"kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q\") pod \"community-operators-qszt8\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.419818 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.420250 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktxp4\" (UniqueName: \"kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.420319 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.497582 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.520963 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.521123 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.521181 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktxp4\" (UniqueName: \"kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.521960 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.522246 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.548845 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktxp4\" (UniqueName: \"kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4\") pod \"certified-operators-txf6q\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:37 crc kubenswrapper[4678]: I1013 13:30:37.649318 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:38 crc kubenswrapper[4678]: I1013 13:30:38.127285 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:38 crc kubenswrapper[4678]: I1013 13:30:38.244713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerStarted","Data":"ebe2504d0a36bdcab34dff6af23dcdf787c78afc540ef32e29d12b961356d94f"} Oct 13 13:30:38 crc kubenswrapper[4678]: I1013 13:30:38.259020 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:39 crc kubenswrapper[4678]: I1013 13:30:39.255600 4678 generic.go:334] "Generic (PLEG): container finished" podID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerID="7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966" exitCode=0 Oct 13 13:30:39 crc kubenswrapper[4678]: I1013 13:30:39.255663 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerDied","Data":"7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966"} Oct 13 13:30:39 crc kubenswrapper[4678]: I1013 13:30:39.256019 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerStarted","Data":"43ae4ff91d8cb963ef13194b041d978a848a8d7a6522d32507f55660bc000dae"} Oct 13 13:30:39 crc kubenswrapper[4678]: I1013 13:30:39.259373 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerID="a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790" exitCode=0 Oct 13 13:30:39 crc kubenswrapper[4678]: I1013 13:30:39.259451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerDied","Data":"a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790"} Oct 13 13:30:40 crc kubenswrapper[4678]: I1013 13:30:40.269052 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerStarted","Data":"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a"} Oct 13 13:30:41 crc kubenswrapper[4678]: I1013 13:30:41.281051 4678 generic.go:334] "Generic (PLEG): container finished" podID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerID="708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671" exitCode=0 Oct 13 13:30:41 crc kubenswrapper[4678]: I1013 13:30:41.281148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerDied","Data":"708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671"} Oct 13 13:30:41 crc kubenswrapper[4678]: I1013 13:30:41.285681 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerID="c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a" exitCode=0 Oct 13 13:30:41 crc kubenswrapper[4678]: I1013 13:30:41.285738 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerDied","Data":"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a"} Oct 13 13:30:42 crc kubenswrapper[4678]: I1013 13:30:42.296154 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerStarted","Data":"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa"} Oct 13 13:30:42 crc kubenswrapper[4678]: I1013 13:30:42.298495 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerStarted","Data":"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77"} Oct 13 13:30:42 crc kubenswrapper[4678]: I1013 13:30:42.332680 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-txf6q" podStartSLOduration=2.848499133 podStartE2EDuration="5.332649762s" podCreationTimestamp="2025-10-13 13:30:37 +0000 UTC" firstStartedPulling="2025-10-13 13:30:39.260640378 +0000 UTC m=+2827.345178272" lastFinishedPulling="2025-10-13 13:30:41.744791017 +0000 UTC m=+2829.829328901" observedRunningTime="2025-10-13 13:30:42.325892426 +0000 UTC m=+2830.410430320" watchObservedRunningTime="2025-10-13 13:30:42.332649762 +0000 UTC m=+2830.417187666" Oct 13 13:30:43 crc kubenswrapper[4678]: I1013 13:30:43.324488 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qszt8" podStartSLOduration=3.7336800500000003 podStartE2EDuration="6.324468589s" podCreationTimestamp="2025-10-13 13:30:37 +0000 UTC" firstStartedPulling="2025-10-13 13:30:39.264098498 +0000 UTC m=+2827.348636382" lastFinishedPulling="2025-10-13 13:30:41.854887037 +0000 UTC m=+2829.939424921" observedRunningTime="2025-10-13 13:30:43.323807432 +0000 UTC m=+2831.408345326" watchObservedRunningTime="2025-10-13 13:30:43.324468589 +0000 UTC m=+2831.409006473" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.498161 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.498661 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.554978 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.650079 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.650128 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:47 crc kubenswrapper[4678]: I1013 13:30:47.695006 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:48 crc kubenswrapper[4678]: I1013 13:30:48.409026 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:48 crc kubenswrapper[4678]: I1013 13:30:48.436850 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:50 crc kubenswrapper[4678]: I1013 13:30:50.541256 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:50 crc kubenswrapper[4678]: I1013 13:30:50.541973 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-txf6q" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="registry-server" containerID="cri-o://dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77" gracePeriod=2 Oct 13 13:30:50 crc kubenswrapper[4678]: I1013 13:30:50.731360 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:50 crc kubenswrapper[4678]: I1013 13:30:50.732067 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qszt8" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="registry-server" containerID="cri-o://9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa" gracePeriod=2 Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.016488 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.161030 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktxp4\" (UniqueName: \"kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4\") pod \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.161149 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content\") pod \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.161359 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities\") pod \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\" (UID: \"f82842ef-05dd-4ad8-aec9-d597be4a7b2d\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.162799 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities" (OuterVolumeSpecName: "utilities") pod "f82842ef-05dd-4ad8-aec9-d597be4a7b2d" (UID: "f82842ef-05dd-4ad8-aec9-d597be4a7b2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.169289 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4" (OuterVolumeSpecName: "kube-api-access-ktxp4") pod "f82842ef-05dd-4ad8-aec9-d597be4a7b2d" (UID: "f82842ef-05dd-4ad8-aec9-d597be4a7b2d"). InnerVolumeSpecName "kube-api-access-ktxp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.209004 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f82842ef-05dd-4ad8-aec9-d597be4a7b2d" (UID: "f82842ef-05dd-4ad8-aec9-d597be4a7b2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.223285 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.264327 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.264391 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktxp4\" (UniqueName: \"kubernetes.io/projected/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-kube-api-access-ktxp4\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.264455 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f82842ef-05dd-4ad8-aec9-d597be4a7b2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.365918 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q994q\" (UniqueName: \"kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q\") pod \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.366145 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content\") pod \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.366181 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities\") pod \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\" (UID: \"e5440eff-b5bb-4182-9350-9d0b7bac47e2\") " Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.367218 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities" (OuterVolumeSpecName: "utilities") pod "e5440eff-b5bb-4182-9350-9d0b7bac47e2" (UID: "e5440eff-b5bb-4182-9350-9d0b7bac47e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.369895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q" (OuterVolumeSpecName: "kube-api-access-q994q") pod "e5440eff-b5bb-4182-9350-9d0b7bac47e2" (UID: "e5440eff-b5bb-4182-9350-9d0b7bac47e2"). InnerVolumeSpecName "kube-api-access-q994q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.389402 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerID="9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa" exitCode=0 Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.389463 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qszt8" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.389490 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerDied","Data":"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa"} Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.389536 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qszt8" event={"ID":"e5440eff-b5bb-4182-9350-9d0b7bac47e2","Type":"ContainerDied","Data":"ebe2504d0a36bdcab34dff6af23dcdf787c78afc540ef32e29d12b961356d94f"} Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.389556 4678 scope.go:117] "RemoveContainer" containerID="9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.393302 4678 generic.go:334] "Generic (PLEG): container finished" podID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerID="dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77" exitCode=0 Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.393340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerDied","Data":"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77"} Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.393369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txf6q" event={"ID":"f82842ef-05dd-4ad8-aec9-d597be4a7b2d","Type":"ContainerDied","Data":"43ae4ff91d8cb963ef13194b041d978a848a8d7a6522d32507f55660bc000dae"} Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.393416 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txf6q" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.410765 4678 scope.go:117] "RemoveContainer" containerID="c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.427818 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5440eff-b5bb-4182-9350-9d0b7bac47e2" (UID: "e5440eff-b5bb-4182-9350-9d0b7bac47e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.437381 4678 scope.go:117] "RemoveContainer" containerID="a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.463377 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469075 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469125 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q994q\" (UniqueName: \"kubernetes.io/projected/e5440eff-b5bb-4182-9350-9d0b7bac47e2-kube-api-access-q994q\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469140 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5440eff-b5bb-4182-9350-9d0b7bac47e2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469348 4678 scope.go:117] "RemoveContainer" containerID="9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.469792 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa\": container with ID starting with 9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa not found: ID does not exist" containerID="9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469836 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa"} err="failed to get container status \"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa\": rpc error: code = NotFound desc = could not find container \"9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa\": container with ID starting with 9854f93fd4f834f16d8471421e2af8a992da3c38b6ff9abedc67a7d13bb936fa not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.469864 4678 scope.go:117] "RemoveContainer" containerID="c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.470387 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a\": container with ID starting with c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a not found: ID does not exist" containerID="c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.470428 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a"} err="failed to get container status \"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a\": rpc error: code = NotFound desc = could not find container \"c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a\": container with ID starting with c60366518097a49da43c75036093e0edb8a303e2a1c7d88bb8484ac8027c9d9a not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.470454 4678 scope.go:117] "RemoveContainer" containerID="a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.470800 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790\": container with ID starting with a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790 not found: ID does not exist" containerID="a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.470835 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790"} err="failed to get container status \"a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790\": rpc error: code = NotFound desc = could not find container \"a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790\": container with ID starting with a9f7ad7a0e3b599fe950a36687abf166a3d609b6219ba20cc700888b5716b790 not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.470863 4678 scope.go:117] "RemoveContainer" containerID="dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.471452 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-txf6q"] Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.486638 4678 scope.go:117] "RemoveContainer" containerID="708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.505474 4678 scope.go:117] "RemoveContainer" containerID="7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.531345 4678 scope.go:117] "RemoveContainer" containerID="dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.531787 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77\": container with ID starting with dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77 not found: ID does not exist" containerID="dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.531830 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77"} err="failed to get container status \"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77\": rpc error: code = NotFound desc = could not find container \"dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77\": container with ID starting with dec44c7f990c274c3497e2c2c2b15ca3a0f76ce5e978f1cbf105012ef95b8d77 not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.531864 4678 scope.go:117] "RemoveContainer" containerID="708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.532357 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671\": container with ID starting with 708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671 not found: ID does not exist" containerID="708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.532396 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671"} err="failed to get container status \"708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671\": rpc error: code = NotFound desc = could not find container \"708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671\": container with ID starting with 708426aa5c8a17f2e03b62c7b8de0fcafd2f83a0aff534bd3769a058e5979671 not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.532421 4678 scope.go:117] "RemoveContainer" containerID="7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966" Oct 13 13:30:51 crc kubenswrapper[4678]: E1013 13:30:51.532868 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966\": container with ID starting with 7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966 not found: ID does not exist" containerID="7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.532918 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966"} err="failed to get container status \"7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966\": rpc error: code = NotFound desc = could not find container \"7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966\": container with ID starting with 7f91e2e6bfa38ad8bc11e4e9eae4e482a3c4f1f74ed5eb1cf432728a9e651966 not found: ID does not exist" Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.752603 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:51 crc kubenswrapper[4678]: I1013 13:30:51.759713 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qszt8"] Oct 13 13:30:52 crc kubenswrapper[4678]: I1013 13:30:52.609005 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" path="/var/lib/kubelet/pods/e5440eff-b5bb-4182-9350-9d0b7bac47e2/volumes" Oct 13 13:30:52 crc kubenswrapper[4678]: I1013 13:30:52.610322 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" path="/var/lib/kubelet/pods/f82842ef-05dd-4ad8-aec9-d597be4a7b2d/volumes" Oct 13 13:30:55 crc kubenswrapper[4678]: I1013 13:30:55.506027 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:30:55 crc kubenswrapper[4678]: I1013 13:30:55.506624 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:30:55 crc kubenswrapper[4678]: I1013 13:30:55.698843 4678 scope.go:117] "RemoveContainer" containerID="599b14dff31b708d53dfdc604db79058284734caa5750a2edfc47cba138d8985" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.506571 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.507183 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.507222 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.507969 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.508015 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" gracePeriod=600 Oct 13 13:31:25 crc kubenswrapper[4678]: E1013 13:31:25.633522 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.687815 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" exitCode=0 Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.687866 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12"} Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.687909 4678 scope.go:117] "RemoveContainer" containerID="b7a54351f39f72e1d54366c7602ebfd8f37b3cb89e8e168b79ef9c09d29cd63b" Oct 13 13:31:25 crc kubenswrapper[4678]: I1013 13:31:25.688602 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:31:25 crc kubenswrapper[4678]: E1013 13:31:25.689105 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:31:37 crc kubenswrapper[4678]: I1013 13:31:37.593572 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:31:37 crc kubenswrapper[4678]: E1013 13:31:37.594406 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:31:50 crc kubenswrapper[4678]: I1013 13:31:50.592971 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:31:50 crc kubenswrapper[4678]: E1013 13:31:50.593706 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:32:02 crc kubenswrapper[4678]: I1013 13:32:02.607035 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:32:02 crc kubenswrapper[4678]: E1013 13:32:02.607828 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:32:15 crc kubenswrapper[4678]: I1013 13:32:15.593144 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:32:15 crc kubenswrapper[4678]: E1013 13:32:15.593933 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:32:28 crc kubenswrapper[4678]: I1013 13:32:28.592009 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:32:28 crc kubenswrapper[4678]: E1013 13:32:28.592760 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:32:42 crc kubenswrapper[4678]: I1013 13:32:42.598541 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:32:42 crc kubenswrapper[4678]: E1013 13:32:42.599513 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:32:55 crc kubenswrapper[4678]: I1013 13:32:55.592763 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:32:55 crc kubenswrapper[4678]: E1013 13:32:55.593681 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:33:07 crc kubenswrapper[4678]: I1013 13:33:07.592556 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:33:07 crc kubenswrapper[4678]: E1013 13:33:07.593363 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:33:20 crc kubenswrapper[4678]: I1013 13:33:20.592641 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:33:20 crc kubenswrapper[4678]: E1013 13:33:20.593918 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:33:31 crc kubenswrapper[4678]: I1013 13:33:31.592880 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:33:31 crc kubenswrapper[4678]: E1013 13:33:31.593672 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:33:46 crc kubenswrapper[4678]: I1013 13:33:46.592428 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:33:46 crc kubenswrapper[4678]: E1013 13:33:46.593274 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:33:59 crc kubenswrapper[4678]: I1013 13:33:59.592289 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:33:59 crc kubenswrapper[4678]: E1013 13:33:59.593000 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:34:10 crc kubenswrapper[4678]: I1013 13:34:10.593711 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:34:10 crc kubenswrapper[4678]: E1013 13:34:10.594695 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:34:25 crc kubenswrapper[4678]: I1013 13:34:25.592835 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:34:25 crc kubenswrapper[4678]: E1013 13:34:25.593634 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:34:40 crc kubenswrapper[4678]: I1013 13:34:40.592193 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:34:40 crc kubenswrapper[4678]: E1013 13:34:40.592927 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:34:54 crc kubenswrapper[4678]: I1013 13:34:54.592878 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:34:54 crc kubenswrapper[4678]: E1013 13:34:54.593742 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:35:05 crc kubenswrapper[4678]: I1013 13:35:05.592942 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:35:05 crc kubenswrapper[4678]: E1013 13:35:05.593759 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:35:20 crc kubenswrapper[4678]: I1013 13:35:20.593094 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:35:20 crc kubenswrapper[4678]: E1013 13:35:20.594906 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:35:33 crc kubenswrapper[4678]: I1013 13:35:33.592779 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:35:33 crc kubenswrapper[4678]: E1013 13:35:33.593486 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:35:48 crc kubenswrapper[4678]: I1013 13:35:48.592711 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:35:48 crc kubenswrapper[4678]: E1013 13:35:48.593599 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:35:59 crc kubenswrapper[4678]: I1013 13:35:59.593334 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:35:59 crc kubenswrapper[4678]: E1013 13:35:59.594368 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:36:12 crc kubenswrapper[4678]: I1013 13:36:12.599517 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:36:12 crc kubenswrapper[4678]: E1013 13:36:12.600985 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:36:24 crc kubenswrapper[4678]: I1013 13:36:24.595816 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:36:24 crc kubenswrapper[4678]: E1013 13:36:24.596510 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:36:36 crc kubenswrapper[4678]: I1013 13:36:36.592868 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:36:37 crc kubenswrapper[4678]: I1013 13:36:37.398090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15"} Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.403847 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407660 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="extract-utilities" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407704 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="extract-utilities" Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407743 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="extract-content" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407753 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="extract-content" Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407774 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="extract-content" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407784 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="extract-content" Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407848 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407858 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407875 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407882 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: E1013 13:37:15.407910 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="extract-utilities" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.407918 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="extract-utilities" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.408203 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5440eff-b5bb-4182-9350-9d0b7bac47e2" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.408225 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f82842ef-05dd-4ad8-aec9-d597be4a7b2d" containerName="registry-server" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.409985 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.426771 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.572179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.572242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gdgb\" (UniqueName: \"kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.572339 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.674423 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.674476 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gdgb\" (UniqueName: \"kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.674528 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.675081 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.675140 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.694696 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gdgb\" (UniqueName: \"kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb\") pod \"redhat-operators-42cs9\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:15 crc kubenswrapper[4678]: I1013 13:37:15.732672 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:16 crc kubenswrapper[4678]: I1013 13:37:16.194593 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:16 crc kubenswrapper[4678]: I1013 13:37:16.748854 4678 generic.go:334] "Generic (PLEG): container finished" podID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerID="d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa" exitCode=0 Oct 13 13:37:16 crc kubenswrapper[4678]: I1013 13:37:16.748911 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerDied","Data":"d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa"} Oct 13 13:37:16 crc kubenswrapper[4678]: I1013 13:37:16.748942 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerStarted","Data":"4ebee85f752ec83570c037a48ac67449779a89e9f023d4218685f2dea3ae0a62"} Oct 13 13:37:16 crc kubenswrapper[4678]: I1013 13:37:16.752309 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:37:18 crc kubenswrapper[4678]: I1013 13:37:18.765205 4678 generic.go:334] "Generic (PLEG): container finished" podID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerID="711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9" exitCode=0 Oct 13 13:37:18 crc kubenswrapper[4678]: I1013 13:37:18.765306 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerDied","Data":"711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9"} Oct 13 13:37:19 crc kubenswrapper[4678]: I1013 13:37:19.777930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerStarted","Data":"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f"} Oct 13 13:37:19 crc kubenswrapper[4678]: I1013 13:37:19.798382 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-42cs9" podStartSLOduration=2.155140211 podStartE2EDuration="4.798362463s" podCreationTimestamp="2025-10-13 13:37:15 +0000 UTC" firstStartedPulling="2025-10-13 13:37:16.75195964 +0000 UTC m=+3224.836497524" lastFinishedPulling="2025-10-13 13:37:19.395181882 +0000 UTC m=+3227.479719776" observedRunningTime="2025-10-13 13:37:19.792917901 +0000 UTC m=+3227.877455785" watchObservedRunningTime="2025-10-13 13:37:19.798362463 +0000 UTC m=+3227.882900347" Oct 13 13:37:25 crc kubenswrapper[4678]: I1013 13:37:25.733461 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:25 crc kubenswrapper[4678]: I1013 13:37:25.734094 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:25 crc kubenswrapper[4678]: I1013 13:37:25.778152 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:25 crc kubenswrapper[4678]: I1013 13:37:25.865287 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:26 crc kubenswrapper[4678]: I1013 13:37:26.009869 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:27 crc kubenswrapper[4678]: I1013 13:37:27.839401 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-42cs9" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="registry-server" containerID="cri-o://9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f" gracePeriod=2 Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.332330 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.527802 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content\") pod \"315fefa6-8f37-47d2-b892-197fe2ff045f\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.528280 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities\") pod \"315fefa6-8f37-47d2-b892-197fe2ff045f\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.528436 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gdgb\" (UniqueName: \"kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb\") pod \"315fefa6-8f37-47d2-b892-197fe2ff045f\" (UID: \"315fefa6-8f37-47d2-b892-197fe2ff045f\") " Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.529009 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities" (OuterVolumeSpecName: "utilities") pod "315fefa6-8f37-47d2-b892-197fe2ff045f" (UID: "315fefa6-8f37-47d2-b892-197fe2ff045f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.534858 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb" (OuterVolumeSpecName: "kube-api-access-5gdgb") pod "315fefa6-8f37-47d2-b892-197fe2ff045f" (UID: "315fefa6-8f37-47d2-b892-197fe2ff045f"). InnerVolumeSpecName "kube-api-access-5gdgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.630895 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.630971 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gdgb\" (UniqueName: \"kubernetes.io/projected/315fefa6-8f37-47d2-b892-197fe2ff045f-kube-api-access-5gdgb\") on node \"crc\" DevicePath \"\"" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.635368 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "315fefa6-8f37-47d2-b892-197fe2ff045f" (UID: "315fefa6-8f37-47d2-b892-197fe2ff045f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.733179 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/315fefa6-8f37-47d2-b892-197fe2ff045f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.849503 4678 generic.go:334] "Generic (PLEG): container finished" podID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerID="9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f" exitCode=0 Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.849569 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-42cs9" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.849599 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerDied","Data":"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f"} Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.850748 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-42cs9" event={"ID":"315fefa6-8f37-47d2-b892-197fe2ff045f","Type":"ContainerDied","Data":"4ebee85f752ec83570c037a48ac67449779a89e9f023d4218685f2dea3ae0a62"} Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.850773 4678 scope.go:117] "RemoveContainer" containerID="9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.887396 4678 scope.go:117] "RemoveContainer" containerID="711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.897791 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.906914 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-42cs9"] Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.916082 4678 scope.go:117] "RemoveContainer" containerID="d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.960812 4678 scope.go:117] "RemoveContainer" containerID="9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f" Oct 13 13:37:28 crc kubenswrapper[4678]: E1013 13:37:28.961371 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f\": container with ID starting with 9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f not found: ID does not exist" containerID="9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.961513 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f"} err="failed to get container status \"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f\": rpc error: code = NotFound desc = could not find container \"9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f\": container with ID starting with 9bd15a4be9a9de053138d9747bb918568f9814813b565db76cbc3cebe979812f not found: ID does not exist" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.961632 4678 scope.go:117] "RemoveContainer" containerID="711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9" Oct 13 13:37:28 crc kubenswrapper[4678]: E1013 13:37:28.962232 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9\": container with ID starting with 711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9 not found: ID does not exist" containerID="711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.962270 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9"} err="failed to get container status \"711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9\": rpc error: code = NotFound desc = could not find container \"711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9\": container with ID starting with 711a9509213e1b8177ba01d027648aece22cc107f337cd00f7de9895ff5e02d9 not found: ID does not exist" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.962300 4678 scope.go:117] "RemoveContainer" containerID="d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa" Oct 13 13:37:28 crc kubenswrapper[4678]: E1013 13:37:28.962574 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa\": container with ID starting with d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa not found: ID does not exist" containerID="d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa" Oct 13 13:37:28 crc kubenswrapper[4678]: I1013 13:37:28.962606 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa"} err="failed to get container status \"d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa\": rpc error: code = NotFound desc = could not find container \"d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa\": container with ID starting with d50696af693aac945d779fef23cc36ae6c249fe27ccb8a1c497416ba1310e3fa not found: ID does not exist" Oct 13 13:37:30 crc kubenswrapper[4678]: I1013 13:37:30.603518 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" path="/var/lib/kubelet/pods/315fefa6-8f37-47d2-b892-197fe2ff045f/volumes" Oct 13 13:38:55 crc kubenswrapper[4678]: I1013 13:38:55.506395 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:38:55 crc kubenswrapper[4678]: I1013 13:38:55.506914 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:39:06 crc kubenswrapper[4678]: I1013 13:39:06.742521 4678 generic.go:334] "Generic (PLEG): container finished" podID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" containerID="69cb42d4e30b310352bfe97fda9019951e7460f234adddfc3b0f056de7397b57" exitCode=0 Oct 13 13:39:06 crc kubenswrapper[4678]: I1013 13:39:06.742597 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a","Type":"ContainerDied","Data":"69cb42d4e30b310352bfe97fda9019951e7460f234adddfc3b0f056de7397b57"} Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.190949 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.371421 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372019 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372087 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372203 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372250 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372293 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372397 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372434 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pdgn\" (UniqueName: \"kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.372492 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key\") pod \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\" (UID: \"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a\") " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.373343 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.373765 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data" (OuterVolumeSpecName: "config-data") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.378083 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn" (OuterVolumeSpecName: "kube-api-access-5pdgn") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "kube-api-access-5pdgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.378178 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.380439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.403805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.420641 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.422487 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.425167 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" (UID: "bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474872 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474901 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-config-data\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474911 4678 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474920 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474930 4678 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474962 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474973 4678 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474984 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pdgn\" (UniqueName: \"kubernetes.io/projected/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-kube-api-access-5pdgn\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.474992 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.494266 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.576511 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.762746 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a","Type":"ContainerDied","Data":"98f8052521704e97feb99d69d89d649d40e40a90c70bdbc5210458c3d700b277"} Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.762789 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98f8052521704e97feb99d69d89d649d40e40a90c70bdbc5210458c3d700b277" Oct 13 13:39:08 crc kubenswrapper[4678]: I1013 13:39:08.762845 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.932508 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 13:39:13 crc kubenswrapper[4678]: E1013 13:39:13.933550 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="extract-utilities" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933570 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="extract-utilities" Oct 13 13:39:13 crc kubenswrapper[4678]: E1013 13:39:13.933606 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="registry-server" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933615 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="registry-server" Oct 13 13:39:13 crc kubenswrapper[4678]: E1013 13:39:13.933635 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" containerName="tempest-tests-tempest-tests-runner" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933645 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" containerName="tempest-tests-tempest-tests-runner" Oct 13 13:39:13 crc kubenswrapper[4678]: E1013 13:39:13.933665 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="extract-content" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933672 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="extract-content" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933871 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a" containerName="tempest-tests-tempest-tests-runner" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.933887 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="315fefa6-8f37-47d2-b892-197fe2ff045f" containerName="registry-server" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.934629 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.938792 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-z78sb" Oct 13 13:39:13 crc kubenswrapper[4678]: I1013 13:39:13.946825 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.087924 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.088445 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4bv\" (UniqueName: \"kubernetes.io/projected/49ee646c-fa4a-4332-92d3-0c70d93e49da-kube-api-access-jl4bv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.190808 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4bv\" (UniqueName: \"kubernetes.io/projected/49ee646c-fa4a-4332-92d3-0c70d93e49da-kube-api-access-jl4bv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.191639 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.192271 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.218277 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4bv\" (UniqueName: \"kubernetes.io/projected/49ee646c-fa4a-4332-92d3-0c70d93e49da-kube-api-access-jl4bv\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.243041 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"49ee646c-fa4a-4332-92d3-0c70d93e49da\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.253201 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.707230 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 13 13:39:14 crc kubenswrapper[4678]: I1013 13:39:14.813710 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"49ee646c-fa4a-4332-92d3-0c70d93e49da","Type":"ContainerStarted","Data":"2f44b6a4b457448f458c8cdb4bdea0399d3cd541e8d98bc3c4e9291758c422e2"} Oct 13 13:39:16 crc kubenswrapper[4678]: I1013 13:39:16.829815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"49ee646c-fa4a-4332-92d3-0c70d93e49da","Type":"ContainerStarted","Data":"c166297c23749cf3e2f7ec588f9af47b60c223c64c4638004383071c61813a51"} Oct 13 13:39:16 crc kubenswrapper[4678]: I1013 13:39:16.842917 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.772406894 podStartE2EDuration="3.842897345s" podCreationTimestamp="2025-10-13 13:39:13 +0000 UTC" firstStartedPulling="2025-10-13 13:39:14.712749356 +0000 UTC m=+3342.797287240" lastFinishedPulling="2025-10-13 13:39:15.783239807 +0000 UTC m=+3343.867777691" observedRunningTime="2025-10-13 13:39:16.841702514 +0000 UTC m=+3344.926240408" watchObservedRunningTime="2025-10-13 13:39:16.842897345 +0000 UTC m=+3344.927435249" Oct 13 13:39:25 crc kubenswrapper[4678]: I1013 13:39:25.505681 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:39:25 crc kubenswrapper[4678]: I1013 13:39:25.506226 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.507422 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xsfjj/must-gather-vs2cg"] Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.509397 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.514733 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xsfjj"/"openshift-service-ca.crt" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.514839 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xsfjj"/"kube-root-ca.crt" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.532633 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xsfjj/must-gather-vs2cg"] Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.662777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.663073 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw8nz\" (UniqueName: \"kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.764711 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw8nz\" (UniqueName: \"kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.765829 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.766293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.787967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw8nz\" (UniqueName: \"kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz\") pod \"must-gather-vs2cg\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:34 crc kubenswrapper[4678]: I1013 13:39:34.825609 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:39:35 crc kubenswrapper[4678]: I1013 13:39:35.293800 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xsfjj/must-gather-vs2cg"] Oct 13 13:39:36 crc kubenswrapper[4678]: I1013 13:39:36.011669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" event={"ID":"d3ba7519-9005-424b-a810-be64a78cf406","Type":"ContainerStarted","Data":"fc76c00d06143613626d06d5cb904ab064e62611fc73fe1bb6336aa5d0f49013"} Oct 13 13:39:40 crc kubenswrapper[4678]: I1013 13:39:40.065913 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" event={"ID":"d3ba7519-9005-424b-a810-be64a78cf406","Type":"ContainerStarted","Data":"6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a"} Oct 13 13:39:40 crc kubenswrapper[4678]: I1013 13:39:40.066407 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" event={"ID":"d3ba7519-9005-424b-a810-be64a78cf406","Type":"ContainerStarted","Data":"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3"} Oct 13 13:39:40 crc kubenswrapper[4678]: I1013 13:39:40.087812 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" podStartSLOduration=2.440913566 podStartE2EDuration="6.08779185s" podCreationTimestamp="2025-10-13 13:39:34 +0000 UTC" firstStartedPulling="2025-10-13 13:39:35.29306545 +0000 UTC m=+3363.377603334" lastFinishedPulling="2025-10-13 13:39:38.939943734 +0000 UTC m=+3367.024481618" observedRunningTime="2025-10-13 13:39:40.082901092 +0000 UTC m=+3368.167438976" watchObservedRunningTime="2025-10-13 13:39:40.08779185 +0000 UTC m=+3368.172329734" Oct 13 13:39:42 crc kubenswrapper[4678]: I1013 13:39:42.930500 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-h7sls"] Oct 13 13:39:42 crc kubenswrapper[4678]: I1013 13:39:42.932565 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:42 crc kubenswrapper[4678]: I1013 13:39:42.935083 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xsfjj"/"default-dockercfg-b8728" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.030692 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.031155 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2knr\" (UniqueName: \"kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.134374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2knr\" (UniqueName: \"kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.134578 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.135097 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.180806 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2knr\" (UniqueName: \"kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr\") pod \"crc-debug-h7sls\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:43 crc kubenswrapper[4678]: I1013 13:39:43.253169 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:39:44 crc kubenswrapper[4678]: I1013 13:39:44.121689 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" event={"ID":"825922a0-7e70-4028-b014-df8d01bfbf35","Type":"ContainerStarted","Data":"5d3253a2504afc1141d661a671525e7f6f7d0643858f503fc3644c1782f947d0"} Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.232486 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" event={"ID":"825922a0-7e70-4028-b014-df8d01bfbf35","Type":"ContainerStarted","Data":"8dc6ab4f1bca1e3158323653a5d195a68924ee8b21479026e308fb1a0cb66db2"} Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.506430 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.506816 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.506859 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.507652 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.507712 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15" gracePeriod=600 Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.558626 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.561667 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.569565 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.702041 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.702279 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.702336 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2c8g\" (UniqueName: \"kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.804638 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.804801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.804860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2c8g\" (UniqueName: \"kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.805456 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.805467 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.826210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2c8g\" (UniqueName: \"kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g\") pod \"redhat-marketplace-mdj9d\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:55 crc kubenswrapper[4678]: I1013 13:39:55.899456 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:39:56 crc kubenswrapper[4678]: I1013 13:39:56.248776 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15" exitCode=0 Oct 13 13:39:56 crc kubenswrapper[4678]: I1013 13:39:56.249761 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15"} Oct 13 13:39:56 crc kubenswrapper[4678]: I1013 13:39:56.249821 4678 scope.go:117] "RemoveContainer" containerID="fbc5cce5b95c500f5647bdedd6dcc29aa83c75e6742fdcaf16c8a3a42aa88f12" Oct 13 13:39:56 crc kubenswrapper[4678]: I1013 13:39:56.271602 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" podStartSLOduration=3.060481702 podStartE2EDuration="14.271583715s" podCreationTimestamp="2025-10-13 13:39:42 +0000 UTC" firstStartedPulling="2025-10-13 13:39:43.297228889 +0000 UTC m=+3371.381766773" lastFinishedPulling="2025-10-13 13:39:54.508330902 +0000 UTC m=+3382.592868786" observedRunningTime="2025-10-13 13:39:56.261481672 +0000 UTC m=+3384.346019556" watchObservedRunningTime="2025-10-13 13:39:56.271583715 +0000 UTC m=+3384.356121599" Oct 13 13:39:56 crc kubenswrapper[4678]: I1013 13:39:56.401008 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:39:56 crc kubenswrapper[4678]: W1013 13:39:56.407328 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79e68e7c_6ad1_4745_9e8d_ceb234b1f25a.slice/crio-5ce1cb9772fd7af5d0aa0fd7651d8749968deaf2a470a5b860dbbd4876b3bec0 WatchSource:0}: Error finding container 5ce1cb9772fd7af5d0aa0fd7651d8749968deaf2a470a5b860dbbd4876b3bec0: Status 404 returned error can't find the container with id 5ce1cb9772fd7af5d0aa0fd7651d8749968deaf2a470a5b860dbbd4876b3bec0 Oct 13 13:39:57 crc kubenswrapper[4678]: I1013 13:39:57.258367 4678 generic.go:334] "Generic (PLEG): container finished" podID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerID="73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74" exitCode=0 Oct 13 13:39:57 crc kubenswrapper[4678]: I1013 13:39:57.258412 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerDied","Data":"73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74"} Oct 13 13:39:57 crc kubenswrapper[4678]: I1013 13:39:57.259027 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerStarted","Data":"5ce1cb9772fd7af5d0aa0fd7651d8749968deaf2a470a5b860dbbd4876b3bec0"} Oct 13 13:39:57 crc kubenswrapper[4678]: I1013 13:39:57.261814 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3"} Oct 13 13:39:58 crc kubenswrapper[4678]: I1013 13:39:58.274427 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerStarted","Data":"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517"} Oct 13 13:39:59 crc kubenswrapper[4678]: I1013 13:39:59.288363 4678 generic.go:334] "Generic (PLEG): container finished" podID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerID="284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517" exitCode=0 Oct 13 13:39:59 crc kubenswrapper[4678]: I1013 13:39:59.288770 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerDied","Data":"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517"} Oct 13 13:40:00 crc kubenswrapper[4678]: I1013 13:40:00.303825 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerStarted","Data":"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6"} Oct 13 13:40:00 crc kubenswrapper[4678]: I1013 13:40:00.329018 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mdj9d" podStartSLOduration=2.834213274 podStartE2EDuration="5.328989631s" podCreationTimestamp="2025-10-13 13:39:55 +0000 UTC" firstStartedPulling="2025-10-13 13:39:57.26079447 +0000 UTC m=+3385.345332354" lastFinishedPulling="2025-10-13 13:39:59.755570827 +0000 UTC m=+3387.840108711" observedRunningTime="2025-10-13 13:40:00.321501435 +0000 UTC m=+3388.406039329" watchObservedRunningTime="2025-10-13 13:40:00.328989631 +0000 UTC m=+3388.413527515" Oct 13 13:40:05 crc kubenswrapper[4678]: I1013 13:40:05.899936 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:05 crc kubenswrapper[4678]: I1013 13:40:05.900638 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:05 crc kubenswrapper[4678]: I1013 13:40:05.950628 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:06 crc kubenswrapper[4678]: I1013 13:40:06.404001 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:06 crc kubenswrapper[4678]: I1013 13:40:06.463940 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.371041 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mdj9d" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="registry-server" containerID="cri-o://7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6" gracePeriod=2 Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.850998 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.948315 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content\") pod \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.948496 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities\") pod \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.948628 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2c8g\" (UniqueName: \"kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g\") pod \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\" (UID: \"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a\") " Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.949512 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities" (OuterVolumeSpecName: "utilities") pod "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" (UID: "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.955198 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g" (OuterVolumeSpecName: "kube-api-access-b2c8g") pod "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" (UID: "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a"). InnerVolumeSpecName "kube-api-access-b2c8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:40:08 crc kubenswrapper[4678]: I1013 13:40:08.962330 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" (UID: "79e68e7c-6ad1-4745-9e8d-ceb234b1f25a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.051098 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.051131 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.051141 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2c8g\" (UniqueName: \"kubernetes.io/projected/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a-kube-api-access-b2c8g\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.384328 4678 generic.go:334] "Generic (PLEG): container finished" podID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerID="7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6" exitCode=0 Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.384396 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerDied","Data":"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6"} Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.384607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mdj9d" event={"ID":"79e68e7c-6ad1-4745-9e8d-ceb234b1f25a","Type":"ContainerDied","Data":"5ce1cb9772fd7af5d0aa0fd7651d8749968deaf2a470a5b860dbbd4876b3bec0"} Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.384643 4678 scope.go:117] "RemoveContainer" containerID="7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.384425 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mdj9d" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.434491 4678 scope.go:117] "RemoveContainer" containerID="284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.443450 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.455160 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mdj9d"] Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.470759 4678 scope.go:117] "RemoveContainer" containerID="73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.556995 4678 scope.go:117] "RemoveContainer" containerID="7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6" Oct 13 13:40:09 crc kubenswrapper[4678]: E1013 13:40:09.557561 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6\": container with ID starting with 7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6 not found: ID does not exist" containerID="7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.557608 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6"} err="failed to get container status \"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6\": rpc error: code = NotFound desc = could not find container \"7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6\": container with ID starting with 7a219b2df1f6f63615e5b4cbe90d6c9782f00c9898f19e5f7511117c7fad90a6 not found: ID does not exist" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.557638 4678 scope.go:117] "RemoveContainer" containerID="284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517" Oct 13 13:40:09 crc kubenswrapper[4678]: E1013 13:40:09.558045 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517\": container with ID starting with 284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517 not found: ID does not exist" containerID="284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.558190 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517"} err="failed to get container status \"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517\": rpc error: code = NotFound desc = could not find container \"284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517\": container with ID starting with 284ccf7e6b2fcb61788e3e0b8afb0579806fad3a718167f5e705699e871cd517 not found: ID does not exist" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.558224 4678 scope.go:117] "RemoveContainer" containerID="73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74" Oct 13 13:40:09 crc kubenswrapper[4678]: E1013 13:40:09.559415 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74\": container with ID starting with 73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74 not found: ID does not exist" containerID="73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74" Oct 13 13:40:09 crc kubenswrapper[4678]: I1013 13:40:09.559454 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74"} err="failed to get container status \"73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74\": rpc error: code = NotFound desc = could not find container \"73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74\": container with ID starting with 73820e56b2bd9a4a0f0e8c4e83632e49f2a143f91bc286b3654e30abfabd3a74 not found: ID does not exist" Oct 13 13:40:10 crc kubenswrapper[4678]: I1013 13:40:10.605314 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" path="/var/lib/kubelet/pods/79e68e7c-6ad1-4745-9e8d-ceb234b1f25a/volumes" Oct 13 13:40:32 crc kubenswrapper[4678]: I1013 13:40:32.565516 4678 generic.go:334] "Generic (PLEG): container finished" podID="825922a0-7e70-4028-b014-df8d01bfbf35" containerID="8dc6ab4f1bca1e3158323653a5d195a68924ee8b21479026e308fb1a0cb66db2" exitCode=0 Oct 13 13:40:32 crc kubenswrapper[4678]: I1013 13:40:32.565606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" event={"ID":"825922a0-7e70-4028-b014-df8d01bfbf35","Type":"ContainerDied","Data":"8dc6ab4f1bca1e3158323653a5d195a68924ee8b21479026e308fb1a0cb66db2"} Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.700036 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.731306 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-h7sls"] Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.737471 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-h7sls"] Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.830703 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host\") pod \"825922a0-7e70-4028-b014-df8d01bfbf35\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.830847 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host" (OuterVolumeSpecName: "host") pod "825922a0-7e70-4028-b014-df8d01bfbf35" (UID: "825922a0-7e70-4028-b014-df8d01bfbf35"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.830905 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2knr\" (UniqueName: \"kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr\") pod \"825922a0-7e70-4028-b014-df8d01bfbf35\" (UID: \"825922a0-7e70-4028-b014-df8d01bfbf35\") " Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.831301 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/825922a0-7e70-4028-b014-df8d01bfbf35-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.836266 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr" (OuterVolumeSpecName: "kube-api-access-v2knr") pod "825922a0-7e70-4028-b014-df8d01bfbf35" (UID: "825922a0-7e70-4028-b014-df8d01bfbf35"). InnerVolumeSpecName "kube-api-access-v2knr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:40:33 crc kubenswrapper[4678]: I1013 13:40:33.933339 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2knr\" (UniqueName: \"kubernetes.io/projected/825922a0-7e70-4028-b014-df8d01bfbf35-kube-api-access-v2knr\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.618602 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-h7sls" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.626472 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="825922a0-7e70-4028-b014-df8d01bfbf35" path="/var/lib/kubelet/pods/825922a0-7e70-4028-b014-df8d01bfbf35/volumes" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.627303 4678 scope.go:117] "RemoveContainer" containerID="8dc6ab4f1bca1e3158323653a5d195a68924ee8b21479026e308fb1a0cb66db2" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.902646 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-ghqjq"] Oct 13 13:40:34 crc kubenswrapper[4678]: E1013 13:40:34.903189 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="registry-server" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903206 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="registry-server" Oct 13 13:40:34 crc kubenswrapper[4678]: E1013 13:40:34.903217 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825922a0-7e70-4028-b014-df8d01bfbf35" containerName="container-00" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903224 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="825922a0-7e70-4028-b014-df8d01bfbf35" containerName="container-00" Oct 13 13:40:34 crc kubenswrapper[4678]: E1013 13:40:34.903243 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="extract-content" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903253 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="extract-content" Oct 13 13:40:34 crc kubenswrapper[4678]: E1013 13:40:34.903265 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="extract-utilities" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903272 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="extract-utilities" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903519 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="79e68e7c-6ad1-4745-9e8d-ceb234b1f25a" containerName="registry-server" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.903535 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="825922a0-7e70-4028-b014-df8d01bfbf35" containerName="container-00" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.904384 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.906573 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xsfjj"/"default-dockercfg-b8728" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.954332 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl6qx\" (UniqueName: \"kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:34 crc kubenswrapper[4678]: I1013 13:40:34.954422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.056036 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl6qx\" (UniqueName: \"kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.056192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.056353 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.078170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl6qx\" (UniqueName: \"kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx\") pod \"crc-debug-ghqjq\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.221470 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.631033 4678 generic.go:334] "Generic (PLEG): container finished" podID="a29ac44b-c543-4b14-9a85-7fb8896a0cee" containerID="f5dbf6f375f65df46685ce9816ecd76edbe7cfff1fcee84230f8eb7285906ef9" exitCode=0 Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.631189 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" event={"ID":"a29ac44b-c543-4b14-9a85-7fb8896a0cee","Type":"ContainerDied","Data":"f5dbf6f375f65df46685ce9816ecd76edbe7cfff1fcee84230f8eb7285906ef9"} Oct 13 13:40:35 crc kubenswrapper[4678]: I1013 13:40:35.631453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" event={"ID":"a29ac44b-c543-4b14-9a85-7fb8896a0cee","Type":"ContainerStarted","Data":"fbf48c1f21e7378efae6fddcefbdeda20ecb0f2897d044f94118b3e443c2e561"} Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.140798 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-ghqjq"] Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.149230 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-ghqjq"] Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.748821 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.788444 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host\") pod \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.788625 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host" (OuterVolumeSpecName: "host") pod "a29ac44b-c543-4b14-9a85-7fb8896a0cee" (UID: "a29ac44b-c543-4b14-9a85-7fb8896a0cee"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.788915 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl6qx\" (UniqueName: \"kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx\") pod \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\" (UID: \"a29ac44b-c543-4b14-9a85-7fb8896a0cee\") " Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.790651 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a29ac44b-c543-4b14-9a85-7fb8896a0cee-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.795177 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx" (OuterVolumeSpecName: "kube-api-access-nl6qx") pod "a29ac44b-c543-4b14-9a85-7fb8896a0cee" (UID: "a29ac44b-c543-4b14-9a85-7fb8896a0cee"). InnerVolumeSpecName "kube-api-access-nl6qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:40:36 crc kubenswrapper[4678]: I1013 13:40:36.893238 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl6qx\" (UniqueName: \"kubernetes.io/projected/a29ac44b-c543-4b14-9a85-7fb8896a0cee-kube-api-access-nl6qx\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.282568 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-z2pjk"] Oct 13 13:40:37 crc kubenswrapper[4678]: E1013 13:40:37.283501 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29ac44b-c543-4b14-9a85-7fb8896a0cee" containerName="container-00" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.283517 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29ac44b-c543-4b14-9a85-7fb8896a0cee" containerName="container-00" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.283809 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29ac44b-c543-4b14-9a85-7fb8896a0cee" containerName="container-00" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.284609 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.402719 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.402851 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnrw4\" (UniqueName: \"kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.505739 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.505897 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.505943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnrw4\" (UniqueName: \"kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.526087 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnrw4\" (UniqueName: \"kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4\") pod \"crc-debug-z2pjk\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.605025 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.682016 4678 scope.go:117] "RemoveContainer" containerID="f5dbf6f375f65df46685ce9816ecd76edbe7cfff1fcee84230f8eb7285906ef9" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.682419 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-ghqjq" Oct 13 13:40:37 crc kubenswrapper[4678]: I1013 13:40:37.707030 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" event={"ID":"c5615e74-e455-4e3a-9df2-2e69fec578c0","Type":"ContainerStarted","Data":"cf528a410723ddb5cb9893fdd1030bdb79ea08805c32bb8d9b14a4073de9bf0c"} Oct 13 13:40:38 crc kubenswrapper[4678]: I1013 13:40:38.606598 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29ac44b-c543-4b14-9a85-7fb8896a0cee" path="/var/lib/kubelet/pods/a29ac44b-c543-4b14-9a85-7fb8896a0cee/volumes" Oct 13 13:40:38 crc kubenswrapper[4678]: I1013 13:40:38.719227 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5615e74-e455-4e3a-9df2-2e69fec578c0" containerID="77b2a00687643c4bc646b5cfe06e1957c901a4acd5bad8199ab74153da61379d" exitCode=0 Oct 13 13:40:38 crc kubenswrapper[4678]: I1013 13:40:38.719271 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" event={"ID":"c5615e74-e455-4e3a-9df2-2e69fec578c0","Type":"ContainerDied","Data":"77b2a00687643c4bc646b5cfe06e1957c901a4acd5bad8199ab74153da61379d"} Oct 13 13:40:38 crc kubenswrapper[4678]: I1013 13:40:38.758916 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-z2pjk"] Oct 13 13:40:38 crc kubenswrapper[4678]: I1013 13:40:38.770352 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xsfjj/crc-debug-z2pjk"] Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.623937 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd59c4d76-26t4f_aef9981e-9e19-4a3a-9f29-a82f4dfdc610/barbican-api/0.log" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.707094 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd59c4d76-26t4f_aef9981e-9e19-4a3a-9f29-a82f4dfdc610/barbican-api-log/0.log" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.829935 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57bd8d65bb-fghbr_8a9e7c5d-3070-4d64-b750-525ec923fc9c/barbican-keystone-listener/0.log" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.840428 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.892951 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57bd8d65bb-fghbr_8a9e7c5d-3070-4d64-b750-525ec923fc9c/barbican-keystone-listener-log/0.log" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.959119 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnrw4\" (UniqueName: \"kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4\") pod \"c5615e74-e455-4e3a-9df2-2e69fec578c0\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.959704 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host\") pod \"c5615e74-e455-4e3a-9df2-2e69fec578c0\" (UID: \"c5615e74-e455-4e3a-9df2-2e69fec578c0\") " Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.960753 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host" (OuterVolumeSpecName: "host") pod "c5615e74-e455-4e3a-9df2-2e69fec578c0" (UID: "c5615e74-e455-4e3a-9df2-2e69fec578c0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:40:39 crc kubenswrapper[4678]: I1013 13:40:39.969528 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4" (OuterVolumeSpecName: "kube-api-access-pnrw4") pod "c5615e74-e455-4e3a-9df2-2e69fec578c0" (UID: "c5615e74-e455-4e3a-9df2-2e69fec578c0"). InnerVolumeSpecName "kube-api-access-pnrw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.055602 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6b7dd7f4b7-jwrwm_b1fd2f17-a805-4985-afba-033768f4d77f/barbican-worker/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.062952 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnrw4\" (UniqueName: \"kubernetes.io/projected/c5615e74-e455-4e3a-9df2-2e69fec578c0-kube-api-access-pnrw4\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.063196 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c5615e74-e455-4e3a-9df2-2e69fec578c0-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.131429 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6b7dd7f4b7-jwrwm_b1fd2f17-a805-4985-afba-033768f4d77f/barbican-worker-log/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.250505 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf_31ca8ec6-826e-4f5b-9e58-160e18674d9f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.350528 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/ceilometer-central-agent/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.384801 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/ceilometer-notification-agent/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.475076 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/proxy-httpd/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.584228 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/sg-core/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.605410 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5615e74-e455-4e3a-9df2-2e69fec578c0" path="/var/lib/kubelet/pods/c5615e74-e455-4e3a-9df2-2e69fec578c0/volumes" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.635821 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8e246acd-cb1e-42d4-935b-53b777639732/cinder-api/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.740159 4678 scope.go:117] "RemoveContainer" containerID="77b2a00687643c4bc646b5cfe06e1957c901a4acd5bad8199ab74153da61379d" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.740192 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/crc-debug-z2pjk" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.783383 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8e246acd-cb1e-42d4-935b-53b777639732/cinder-api-log/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.897920 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_af635b84-5ac7-42ee-ac5c-55a2c5e20932/probe/0.log" Oct 13 13:40:40 crc kubenswrapper[4678]: I1013 13:40:40.981612 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_af635b84-5ac7-42ee-ac5c-55a2c5e20932/cinder-scheduler/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.133675 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z_e0fd32ad-0b5d-4287-a5db-79c362e818e6/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.194884 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz_567d50a7-a8dd-4bdb-82dc-8bee34a93d25/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.375221 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd_0dacde92-4685-4950-8219-d634cbd0b1cc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.429121 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/init/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.579022 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/init/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.659884 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/dnsmasq-dns/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.667701 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-hn894_48c09433-04bb-4cd0-8bfd-72606ebb285a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.862034 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1b7bcad6-33d5-425e-aeda-70bf6518c557/glance-httpd/0.log" Oct 13 13:40:41 crc kubenswrapper[4678]: I1013 13:40:41.863937 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1b7bcad6-33d5-425e-aeda-70bf6518c557/glance-log/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.077676 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0b4628b6-c114-4448-b559-79861b376761/glance-httpd/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.097038 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0b4628b6-c114-4448-b559-79861b376761/glance-log/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.373320 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94756fb8d-9h8cs_5bc24a68-de18-47d6-9573-67e978a05b05/horizon/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.473636 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6_3608a541-f8d9-4931-a7d1-cbf90f6c505f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.623377 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-ssndt_84dfdaf8-938e-4eca-af31-79b85a483599/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.627164 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94756fb8d-9h8cs_5bc24a68-de18-47d6-9573-67e978a05b05/horizon-log/0.log" Oct 13 13:40:42 crc kubenswrapper[4678]: I1013 13:40:42.939262 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29339341-xm4f5_2a462e27-6016-43fd-8989-d97b06e1ea3c/keystone-cron/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.016918 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b4bffbf65-fc8vg_f6470c70-b6b8-456a-9ca0-5bc9c9122674/keystone-api/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.133064 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_20a83011-dd00-465e-97cc-0bc354fec144/kube-state-metrics/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.305623 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m62j6_08cf88a5-76f1-455d-a838-9d8e37902c6e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.824137 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dcbfc8897-x7vwx_cc052259-a50b-4d53-ba5a-bbc8458aaa97/neutron-httpd/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.830936 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dcbfc8897-x7vwx_cc052259-a50b-4d53-ba5a-bbc8458aaa97/neutron-api/0.log" Oct 13 13:40:43 crc kubenswrapper[4678]: I1013 13:40:43.894868 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6_8a30e130-e610-4241-beb7-0a5d88fda83b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:44 crc kubenswrapper[4678]: I1013 13:40:44.414888 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6c38d115-b798-4e2c-986c-c9541832f1bb/nova-api-log/0.log" Oct 13 13:40:44 crc kubenswrapper[4678]: I1013 13:40:44.574321 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0f4da22c-d9e7-4f23-af49-a6ed4551b4b8/nova-cell0-conductor-conductor/0.log" Oct 13 13:40:44 crc kubenswrapper[4678]: I1013 13:40:44.728684 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_902f91dc-30b7-4350-9d03-9517e3b911ae/nova-cell1-conductor-conductor/0.log" Oct 13 13:40:44 crc kubenswrapper[4678]: I1013 13:40:44.730737 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6c38d115-b798-4e2c-986c-c9541832f1bb/nova-api-api/0.log" Oct 13 13:40:44 crc kubenswrapper[4678]: I1013 13:40:44.853640 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5b4c17c9-33f5-406f-9e2f-2fa0eef6171d/nova-cell1-novncproxy-novncproxy/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.009517 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-w28hm_204374ee-a5e4-4a02-bada-f57883e5c571/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.194651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0b7bb30d-3831-467e-8067-67626163b30a/nova-metadata-log/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.397001 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ac735c75-e8f1-402a-a7d7-d213a9badd88/nova-scheduler-scheduler/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.530860 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/mysql-bootstrap/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.731093 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/galera/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.755916 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/mysql-bootstrap/0.log" Oct 13 13:40:45 crc kubenswrapper[4678]: I1013 13:40:45.956250 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/mysql-bootstrap/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.144717 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/mysql-bootstrap/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.181423 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/galera/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.330828 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b10ccdc-9b5e-4093-8a72-5541ad94e2f3/openstackclient/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.442799 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-dv2nq_2e724ee3-a6c7-4b75-9c60-3fe2f233b96b/ovn-controller/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.447189 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0b7bb30d-3831-467e-8067-67626163b30a/nova-metadata-metadata/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.672701 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-67jfd_e2f127fd-3a0a-44c7-9849-de6e42149fd7/openstack-network-exporter/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.761424 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server-init/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.898444 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server-init/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.914128 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovs-vswitchd/0.log" Oct 13 13:40:46 crc kubenswrapper[4678]: I1013 13:40:46.977398 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.154660 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lvc85_416f5c49-3fd6-4789-924a-4e0ba0279fa5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.224274 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_030132a0-ddd0-4001-9216-9a2962cb9f1c/openstack-network-exporter/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.361388 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_030132a0-ddd0-4001-9216-9a2962cb9f1c/ovn-northd/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.436094 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b1da318b-1924-4ce5-9a9a-a1b483c0ffb0/openstack-network-exporter/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.463190 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b1da318b-1924-4ce5-9a9a-a1b483c0ffb0/ovsdbserver-nb/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.571101 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_352ce882-938a-4fe3-b4ce-7e01a5a750e6/openstack-network-exporter/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.700305 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_352ce882-938a-4fe3-b4ce-7e01a5a750e6/ovsdbserver-sb/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.952855 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-694bd84568-p2blx_776b5cb1-0247-44a1-b63b-cffb9a30e588/placement-log/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.954947 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-694bd84568-p2blx_776b5cb1-0247-44a1-b63b-cffb9a30e588/placement-api/0.log" Oct 13 13:40:47 crc kubenswrapper[4678]: I1013 13:40:47.983514 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/setup-container/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.188494 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/setup-container/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.255201 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/rabbitmq/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.261341 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/setup-container/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.538162 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/setup-container/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.545707 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/rabbitmq/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.573515 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv_b8dc18d2-df42-4546-981a-cfe13dcd879f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.824376 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s9vpt_26b2ddaa-5eaa-48c5-947b-507b97253a67/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:48 crc kubenswrapper[4678]: I1013 13:40:48.834166 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw_d4c81445-8564-47a3-ab04-db591cbf1a86/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.007034 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pbbsm_4db4bd12-e81c-4bcc-8f80-ed20807d5856/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.092501 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2p9wv_78c20a4a-fd2b-4fa7-bf79-4ae7559b312e/ssh-known-hosts-edpm-deployment/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.360176 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f9fcfccd7-n448c_6d9ab7cf-ca73-4d0f-9338-68536f651533/proxy-server/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.366150 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f9fcfccd7-n448c_6d9ab7cf-ca73-4d0f-9338-68536f651533/proxy-httpd/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.467028 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6x2jw_07587cb9-f0d2-41d5-a596-7650a30f9cd1/swift-ring-rebalance/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.627797 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-auditor/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.690179 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-reaper/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.806476 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-replicator/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.885326 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-server/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.903518 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-auditor/0.log" Oct 13 13:40:49 crc kubenswrapper[4678]: I1013 13:40:49.930020 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-replicator/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.019554 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-server/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.125170 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-updater/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.181433 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-expirer/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.191313 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-auditor/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.262632 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-replicator/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.328851 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-server/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.372589 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-updater/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.412173 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/rsync/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.508734 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/swift-recon-cron/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.634279 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x56k2_d7940e67-463a-4cc5-812e-0bde560d6eda/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.720673 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a/tempest-tests-tempest-tests-runner/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.869227 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_49ee646c-fa4a-4332-92d3-0c70d93e49da/test-operator-logs-container/0.log" Oct 13 13:40:50 crc kubenswrapper[4678]: I1013 13:40:50.987145 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-sqqln_5fc607e5-2f31-42b9-ba24-ba02b1516a75/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:40:59 crc kubenswrapper[4678]: I1013 13:40:59.321593 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_911b721e-cee9-45db-ae01-fb84bb431847/memcached/0.log" Oct 13 13:41:14 crc kubenswrapper[4678]: I1013 13:41:14.503361 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:41:14 crc kubenswrapper[4678]: I1013 13:41:14.794257 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:41:14 crc kubenswrapper[4678]: I1013 13:41:14.795592 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:41:14 crc kubenswrapper[4678]: I1013 13:41:14.809355 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.029922 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.035626 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.067296 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/extract/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.232576 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-srpr5_efab1e79-db75-4c28-9046-f5fb22a49c76/kube-rbac-proxy/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.295537 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-srpr5_efab1e79-db75-4c28-9046-f5fb22a49c76/manager/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.355934 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-hcblt_7809e838-a0e6-4bc7-828f-bc6aaa4ebca5/kube-rbac-proxy/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.543136 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ltlvn_7a378ef8-b17b-465d-aabb-e289cd9719cf/kube-rbac-proxy/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.582619 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-hcblt_7809e838-a0e6-4bc7-828f-bc6aaa4ebca5/manager/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.592603 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ltlvn_7a378ef8-b17b-465d-aabb-e289cd9719cf/manager/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.740019 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-wf2p9_93f6cd1f-8a49-4a2b-927a-0c4a5654891d/kube-rbac-proxy/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.873578 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-wf2p9_93f6cd1f-8a49-4a2b-927a-0c4a5654891d/manager/0.log" Oct 13 13:41:15 crc kubenswrapper[4678]: I1013 13:41:15.983566 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-c74gx_2925bcf7-ade2-4da7-a99e-3d187eafd11a/manager/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.005328 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-c74gx_2925bcf7-ade2-4da7-a99e-3d187eafd11a/kube-rbac-proxy/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.093099 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-xg5q2_421e19af-df3c-4736-a6d1-ba43a6f8856d/kube-rbac-proxy/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.294406 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-xb7kd_cc1d3202-a214-4cbb-a50a-7d3ecb20409f/kube-rbac-proxy/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.304702 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-xg5q2_421e19af-df3c-4736-a6d1-ba43a6f8856d/manager/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.480649 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-xb7kd_cc1d3202-a214-4cbb-a50a-7d3ecb20409f/manager/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.492877 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vfc96_11cc73f1-36f5-4221-9b67-fb4e7611ae65/kube-rbac-proxy/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.608096 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vfc96_11cc73f1-36f5-4221-9b67-fb4e7611ae65/manager/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.708652 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-qwthm_c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1/kube-rbac-proxy/0.log" Oct 13 13:41:16 crc kubenswrapper[4678]: I1013 13:41:16.814499 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-qwthm_c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.156567 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-64wl2_872c1819-1912-4171-8e47-35843008daaf/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.178566 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-64wl2_872c1819-1912-4171-8e47-35843008daaf/kube-rbac-proxy/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.374597 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-796dd_781f6537-803f-476e-a741-2030455c777f/kube-rbac-proxy/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.411409 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-796dd_781f6537-803f-476e-a741-2030455c777f/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.484770 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-f4f4z_0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a/kube-rbac-proxy/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.630693 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-f4f4z_0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.642091 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-hgmrn_d89268fa-c74a-49e5-9988-e6a9fb33d5c4/kube-rbac-proxy/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.845440 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-hgmrn_d89268fa-c74a-49e5-9988-e6a9fb33d5c4/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.886787 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-h9vf6_b31e9b57-89fe-42a7-ad38-3ae06224ae58/manager/0.log" Oct 13 13:41:17 crc kubenswrapper[4678]: I1013 13:41:17.906201 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-h9vf6_b31e9b57-89fe-42a7-ad38-3ae06224ae58/kube-rbac-proxy/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.074871 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z_a2deeac1-517e-4e4f-974a-1f0267868448/kube-rbac-proxy/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.129155 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z_a2deeac1-517e-4e4f-974a-1f0267868448/manager/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.266734 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d8c8d84c-lsmxm_3137ba4b-0e52-456e-9951-08989c113034/kube-rbac-proxy/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.390118 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-766c4fbc4c-ddcpv_e6f72e4a-d252-4053-9665-db9af69e697e/kube-rbac-proxy/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.595068 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4fn9q_dadbd9ec-7f53-4555-951c-2d090c017417/registry-server/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.615811 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-766c4fbc4c-ddcpv_e6f72e4a-d252-4053-9665-db9af69e697e/operator/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.813983 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-5rt5t_953d626f-b1c6-492b-8473-52da7c6edc3f/kube-rbac-proxy/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.868336 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-5rt5t_953d626f-b1c6-492b-8473-52da7c6edc3f/manager/0.log" Oct 13 13:41:18 crc kubenswrapper[4678]: I1013 13:41:18.929638 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-wvf4s_b2063d32-c006-427e-b9a6-84007b0c1d22/kube-rbac-proxy/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.097372 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-wvf4s_b2063d32-c006-427e-b9a6-84007b0c1d22/manager/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.102745 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz_6083ce11-ee7f-4bbc-b5ff-c2816dfe6300/operator/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.336946 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-v5b9q_2e600274-90bb-46a9-b41a-f642ed284900/manager/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.354027 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-v5b9q_2e600274-90bb-46a9-b41a-f642ed284900/kube-rbac-proxy/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.376676 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qzgc8_b7188e6b-9e98-477e-bbe8-778da7f9bda1/kube-rbac-proxy/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.467279 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d8c8d84c-lsmxm_3137ba4b-0e52-456e-9951-08989c113034/manager/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.577517 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-46shf_83789dd0-5367-44d6-8cf9-0f4bf92280b1/kube-rbac-proxy/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.613402 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qzgc8_b7188e6b-9e98-477e-bbe8-778da7f9bda1/manager/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.664087 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-46shf_83789dd0-5367-44d6-8cf9-0f4bf92280b1/manager/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.765446 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-8dkj8_ec1fafb9-7b71-45b3-ade5-6df84a9fa568/kube-rbac-proxy/0.log" Oct 13 13:41:19 crc kubenswrapper[4678]: I1013 13:41:19.810434 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-8dkj8_ec1fafb9-7b71-45b3-ade5-6df84a9fa568/manager/0.log" Oct 13 13:41:35 crc kubenswrapper[4678]: I1013 13:41:35.351350 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-m99jv_ad4e211d-f191-4262-9e70-d1f4fcaddeff/control-plane-machine-set-operator/0.log" Oct 13 13:41:35 crc kubenswrapper[4678]: I1013 13:41:35.557434 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dkxs9_3da9a359-7295-4aa0-95e0-070be29e5eb9/kube-rbac-proxy/0.log" Oct 13 13:41:35 crc kubenswrapper[4678]: I1013 13:41:35.596856 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dkxs9_3da9a359-7295-4aa0-95e0-070be29e5eb9/machine-api-operator/0.log" Oct 13 13:41:48 crc kubenswrapper[4678]: I1013 13:41:48.088229 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vw4gz_e0a73006-9a7f-41d4-a99c-d749fdfad405/cert-manager-controller/0.log" Oct 13 13:41:48 crc kubenswrapper[4678]: I1013 13:41:48.359852 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-l8c98_2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1/cert-manager-cainjector/0.log" Oct 13 13:41:48 crc kubenswrapper[4678]: I1013 13:41:48.415476 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-pczp4_2d01b76f-171e-4e27-a020-3e43c2405b19/cert-manager-webhook/0.log" Oct 13 13:41:59 crc kubenswrapper[4678]: I1013 13:41:59.676413 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-c5csj_e3884e53-e229-44e5-8d8c-42fcc654f657/nmstate-console-plugin/0.log" Oct 13 13:41:59 crc kubenswrapper[4678]: I1013 13:41:59.904590 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rtdjd_5439ee39-b0d6-43fc-92aa-f5240c460afc/nmstate-handler/0.log" Oct 13 13:41:59 crc kubenswrapper[4678]: I1013 13:41:59.927777 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-nws2x_00b6c406-a244-403d-9ca0-2406de2da1f5/kube-rbac-proxy/0.log" Oct 13 13:41:59 crc kubenswrapper[4678]: I1013 13:41:59.988242 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-nws2x_00b6c406-a244-403d-9ca0-2406de2da1f5/nmstate-metrics/0.log" Oct 13 13:42:00 crc kubenswrapper[4678]: I1013 13:42:00.154085 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hk95b_0a00ef2f-1663-499a-97a7-1b14e7d7f3c1/nmstate-operator/0.log" Oct 13 13:42:00 crc kubenswrapper[4678]: I1013 13:42:00.191905 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-tbt2v_ae486b1a-e0b6-45ff-91f2-6bb967b88436/nmstate-webhook/0.log" Oct 13 13:42:12 crc kubenswrapper[4678]: I1013 13:42:12.993409 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-x9tg5_ae7e1bbe-3d07-40db-991b-797c25efd443/kube-rbac-proxy/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.127113 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-x9tg5_ae7e1bbe-3d07-40db-991b-797c25efd443/controller/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.214651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.409100 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.431004 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.472725 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.476944 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.638980 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.639970 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.656019 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:42:13 crc kubenswrapper[4678]: I1013 13:42:13.674651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.004702 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.019704 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/controller/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.039919 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.040981 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.197020 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/frr-metrics/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.295233 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/kube-rbac-proxy-frr/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.296721 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/kube-rbac-proxy/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.505974 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/reloader/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.524817 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-xp775_61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac/frr-k8s-webhook-server/0.log" Oct 13 13:42:14 crc kubenswrapper[4678]: I1013 13:42:14.710928 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-797dff698d-tbgv4_51bd00ff-af4f-4405-9f25-20b4d22e7020/manager/0.log" Oct 13 13:42:15 crc kubenswrapper[4678]: I1013 13:42:15.017392 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6545fd778d-8xqwq_0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5/webhook-server/0.log" Oct 13 13:42:15 crc kubenswrapper[4678]: I1013 13:42:15.020439 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mbsh9_f0f04693-55fb-40f0-8f97-c91d12a02509/kube-rbac-proxy/0.log" Oct 13 13:42:15 crc kubenswrapper[4678]: I1013 13:42:15.494675 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/frr/0.log" Oct 13 13:42:15 crc kubenswrapper[4678]: I1013 13:42:15.603021 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mbsh9_f0f04693-55fb-40f0-8f97-c91d12a02509/speaker/0.log" Oct 13 13:42:25 crc kubenswrapper[4678]: I1013 13:42:25.506313 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:42:25 crc kubenswrapper[4678]: I1013 13:42:25.506982 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:42:27 crc kubenswrapper[4678]: I1013 13:42:27.743914 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.106961 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.112485 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.174290 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.305693 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.359479 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.370635 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/extract/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.489874 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.623520 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.647411 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.653760 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.834171 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:42:28 crc kubenswrapper[4678]: I1013 13:42:28.844676 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.071806 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.227445 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/registry-server/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.254260 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.279799 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.316888 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.526876 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.541553 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.770379 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.866691 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/registry-server/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.963394 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.974457 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:42:29 crc kubenswrapper[4678]: I1013 13:42:29.998697 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.164545 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.168523 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.205887 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/extract/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.339976 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9j9mq_2c14e376-1a85-43dc-a3d1-66e37a54fd22/marketplace-operator/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.395262 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.606350 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.635770 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.663495 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.795427 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.811898 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.938831 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/registry-server/0.log" Oct 13 13:42:30 crc kubenswrapper[4678]: I1013 13:42:30.977142 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.138911 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.176808 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.204840 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.364748 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.394466 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:42:31 crc kubenswrapper[4678]: I1013 13:42:31.977168 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/registry-server/0.log" Oct 13 13:42:48 crc kubenswrapper[4678]: E1013 13:42:48.835692 4678 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.30:48596->38.102.83.30:36753: write tcp 38.102.83.30:48596->38.102.83.30:36753: write: broken pipe Oct 13 13:42:55 crc kubenswrapper[4678]: I1013 13:42:55.505771 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:42:55 crc kubenswrapper[4678]: I1013 13:42:55.506446 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:43:25 crc kubenswrapper[4678]: I1013 13:43:25.505813 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:43:25 crc kubenswrapper[4678]: I1013 13:43:25.506458 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 13 13:43:25 crc kubenswrapper[4678]: I1013 13:43:25.506523 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" Oct 13 13:43:25 crc kubenswrapper[4678]: I1013 13:43:25.507374 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3"} pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 13 13:43:25 crc kubenswrapper[4678]: I1013 13:43:25.507432 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" containerID="cri-o://474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" gracePeriod=600 Oct 13 13:43:25 crc kubenswrapper[4678]: E1013 13:43:25.636088 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:43:26 crc kubenswrapper[4678]: I1013 13:43:26.270770 4678 generic.go:334] "Generic (PLEG): container finished" podID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" exitCode=0 Oct 13 13:43:26 crc kubenswrapper[4678]: I1013 13:43:26.270813 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerDied","Data":"474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3"} Oct 13 13:43:26 crc kubenswrapper[4678]: I1013 13:43:26.270845 4678 scope.go:117] "RemoveContainer" containerID="60241e0031a2fe614e99026c035a2c10d6fe90d72648f5276353498534446d15" Oct 13 13:43:26 crc kubenswrapper[4678]: I1013 13:43:26.272738 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:43:26 crc kubenswrapper[4678]: E1013 13:43:26.273204 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:43:38 crc kubenswrapper[4678]: I1013 13:43:38.593044 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:43:38 crc kubenswrapper[4678]: E1013 13:43:38.594778 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:43:53 crc kubenswrapper[4678]: I1013 13:43:53.593878 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:43:53 crc kubenswrapper[4678]: E1013 13:43:53.594920 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:44:05 crc kubenswrapper[4678]: I1013 13:44:05.592959 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:44:05 crc kubenswrapper[4678]: E1013 13:44:05.594560 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:44:08 crc kubenswrapper[4678]: I1013 13:44:08.722224 4678 generic.go:334] "Generic (PLEG): container finished" podID="d3ba7519-9005-424b-a810-be64a78cf406" containerID="c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3" exitCode=0 Oct 13 13:44:08 crc kubenswrapper[4678]: I1013 13:44:08.722349 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" event={"ID":"d3ba7519-9005-424b-a810-be64a78cf406","Type":"ContainerDied","Data":"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3"} Oct 13 13:44:08 crc kubenswrapper[4678]: I1013 13:44:08.724132 4678 scope.go:117] "RemoveContainer" containerID="c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3" Oct 13 13:44:09 crc kubenswrapper[4678]: I1013 13:44:09.261519 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xsfjj_must-gather-vs2cg_d3ba7519-9005-424b-a810-be64a78cf406/gather/0.log" Oct 13 13:44:16 crc kubenswrapper[4678]: I1013 13:44:16.993306 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xsfjj/must-gather-vs2cg"] Oct 13 13:44:16 crc kubenswrapper[4678]: I1013 13:44:16.994186 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="copy" containerID="cri-o://6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a" gracePeriod=2 Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.006245 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xsfjj/must-gather-vs2cg"] Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.423878 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xsfjj_must-gather-vs2cg_d3ba7519-9005-424b-a810-be64a78cf406/copy/0.log" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.424688 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.481858 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw8nz\" (UniqueName: \"kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz\") pod \"d3ba7519-9005-424b-a810-be64a78cf406\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.481911 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output\") pod \"d3ba7519-9005-424b-a810-be64a78cf406\" (UID: \"d3ba7519-9005-424b-a810-be64a78cf406\") " Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.488020 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz" (OuterVolumeSpecName: "kube-api-access-bw8nz") pod "d3ba7519-9005-424b-a810-be64a78cf406" (UID: "d3ba7519-9005-424b-a810-be64a78cf406"). InnerVolumeSpecName "kube-api-access-bw8nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.585410 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw8nz\" (UniqueName: \"kubernetes.io/projected/d3ba7519-9005-424b-a810-be64a78cf406-kube-api-access-bw8nz\") on node \"crc\" DevicePath \"\"" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.654297 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d3ba7519-9005-424b-a810-be64a78cf406" (UID: "d3ba7519-9005-424b-a810-be64a78cf406"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.686869 4678 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d3ba7519-9005-424b-a810-be64a78cf406-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.802384 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xsfjj_must-gather-vs2cg_d3ba7519-9005-424b-a810-be64a78cf406/copy/0.log" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.802906 4678 generic.go:334] "Generic (PLEG): container finished" podID="d3ba7519-9005-424b-a810-be64a78cf406" containerID="6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a" exitCode=143 Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.802963 4678 scope.go:117] "RemoveContainer" containerID="6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.802966 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xsfjj/must-gather-vs2cg" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.825813 4678 scope.go:117] "RemoveContainer" containerID="c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.886758 4678 scope.go:117] "RemoveContainer" containerID="6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a" Oct 13 13:44:17 crc kubenswrapper[4678]: E1013 13:44:17.888477 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a\": container with ID starting with 6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a not found: ID does not exist" containerID="6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.888540 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a"} err="failed to get container status \"6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a\": rpc error: code = NotFound desc = could not find container \"6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a\": container with ID starting with 6f15ff96bd77a514ea3b3789f95ca213d4141497d6ccdac7b2bc3f0a6700727a not found: ID does not exist" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.888579 4678 scope.go:117] "RemoveContainer" containerID="c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3" Oct 13 13:44:17 crc kubenswrapper[4678]: E1013 13:44:17.888920 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3\": container with ID starting with c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3 not found: ID does not exist" containerID="c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3" Oct 13 13:44:17 crc kubenswrapper[4678]: I1013 13:44:17.888960 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3"} err="failed to get container status \"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3\": rpc error: code = NotFound desc = could not find container \"c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3\": container with ID starting with c79c9494d8a70b640469223b9d5872e4d75d07ba0ad16e58f5ac0bbf5470f3b3 not found: ID does not exist" Oct 13 13:44:18 crc kubenswrapper[4678]: I1013 13:44:18.613951 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3ba7519-9005-424b-a810-be64a78cf406" path="/var/lib/kubelet/pods/d3ba7519-9005-424b-a810-be64a78cf406/volumes" Oct 13 13:44:20 crc kubenswrapper[4678]: I1013 13:44:20.593045 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:44:20 crc kubenswrapper[4678]: E1013 13:44:20.593847 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:44:35 crc kubenswrapper[4678]: I1013 13:44:35.592943 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:44:35 crc kubenswrapper[4678]: E1013 13:44:35.593799 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:44:46 crc kubenswrapper[4678]: I1013 13:44:46.592831 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:44:46 crc kubenswrapper[4678]: E1013 13:44:46.593914 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:44:57 crc kubenswrapper[4678]: I1013 13:44:57.592522 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:44:57 crc kubenswrapper[4678]: E1013 13:44:57.593340 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.189278 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks"] Oct 13 13:45:00 crc kubenswrapper[4678]: E1013 13:45:00.189942 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5615e74-e455-4e3a-9df2-2e69fec578c0" containerName="container-00" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.189955 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5615e74-e455-4e3a-9df2-2e69fec578c0" containerName="container-00" Oct 13 13:45:00 crc kubenswrapper[4678]: E1013 13:45:00.189989 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="gather" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.189994 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="gather" Oct 13 13:45:00 crc kubenswrapper[4678]: E1013 13:45:00.190011 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="copy" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.190018 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="copy" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.190208 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5615e74-e455-4e3a-9df2-2e69fec578c0" containerName="container-00" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.190227 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="gather" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.190238 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3ba7519-9005-424b-a810-be64a78cf406" containerName="copy" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.190961 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.193904 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.193985 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.206689 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks"] Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.324141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqm29\" (UniqueName: \"kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.324246 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.324274 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.426318 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqm29\" (UniqueName: \"kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.426429 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.426458 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.427391 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.433574 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.447366 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqm29\" (UniqueName: \"kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29\") pod \"collect-profiles-29339385-m9vks\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.526478 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:00 crc kubenswrapper[4678]: I1013 13:45:00.981319 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks"] Oct 13 13:45:01 crc kubenswrapper[4678]: I1013 13:45:01.253382 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" event={"ID":"cf7d1823-af07-4cea-9059-23707cbffcb1","Type":"ContainerStarted","Data":"013b9b067d344d817e94171089ffb69560493ca24e4290d40fbe782207dcda7b"} Oct 13 13:45:01 crc kubenswrapper[4678]: I1013 13:45:01.253701 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" event={"ID":"cf7d1823-af07-4cea-9059-23707cbffcb1","Type":"ContainerStarted","Data":"165237fb87d7d438690faa5d9c1751379b9212a8ba4a27478f5e3e3bad53a52c"} Oct 13 13:45:01 crc kubenswrapper[4678]: I1013 13:45:01.281272 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" podStartSLOduration=1.28124997 podStartE2EDuration="1.28124997s" podCreationTimestamp="2025-10-13 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:45:01.270773017 +0000 UTC m=+3689.355310901" watchObservedRunningTime="2025-10-13 13:45:01.28124997 +0000 UTC m=+3689.365787854" Oct 13 13:45:02 crc kubenswrapper[4678]: I1013 13:45:02.269233 4678 generic.go:334] "Generic (PLEG): container finished" podID="cf7d1823-af07-4cea-9059-23707cbffcb1" containerID="013b9b067d344d817e94171089ffb69560493ca24e4290d40fbe782207dcda7b" exitCode=0 Oct 13 13:45:02 crc kubenswrapper[4678]: I1013 13:45:02.269456 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" event={"ID":"cf7d1823-af07-4cea-9059-23707cbffcb1","Type":"ContainerDied","Data":"013b9b067d344d817e94171089ffb69560493ca24e4290d40fbe782207dcda7b"} Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.583040 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.706378 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqm29\" (UniqueName: \"kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29\") pod \"cf7d1823-af07-4cea-9059-23707cbffcb1\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.706496 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume\") pod \"cf7d1823-af07-4cea-9059-23707cbffcb1\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.706558 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume\") pod \"cf7d1823-af07-4cea-9059-23707cbffcb1\" (UID: \"cf7d1823-af07-4cea-9059-23707cbffcb1\") " Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.707473 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume" (OuterVolumeSpecName: "config-volume") pod "cf7d1823-af07-4cea-9059-23707cbffcb1" (UID: "cf7d1823-af07-4cea-9059-23707cbffcb1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.711790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29" (OuterVolumeSpecName: "kube-api-access-lqm29") pod "cf7d1823-af07-4cea-9059-23707cbffcb1" (UID: "cf7d1823-af07-4cea-9059-23707cbffcb1"). InnerVolumeSpecName "kube-api-access-lqm29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.712215 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cf7d1823-af07-4cea-9059-23707cbffcb1" (UID: "cf7d1823-af07-4cea-9059-23707cbffcb1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.808893 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cf7d1823-af07-4cea-9059-23707cbffcb1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.808939 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqm29\" (UniqueName: \"kubernetes.io/projected/cf7d1823-af07-4cea-9059-23707cbffcb1-kube-api-access-lqm29\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:03 crc kubenswrapper[4678]: I1013 13:45:03.808956 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cf7d1823-af07-4cea-9059-23707cbffcb1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.290955 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" event={"ID":"cf7d1823-af07-4cea-9059-23707cbffcb1","Type":"ContainerDied","Data":"165237fb87d7d438690faa5d9c1751379b9212a8ba4a27478f5e3e3bad53a52c"} Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.291325 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="165237fb87d7d438690faa5d9c1751379b9212a8ba4a27478f5e3e3bad53a52c" Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.291010 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29339385-m9vks" Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.355511 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4"] Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.366266 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29339340-mzxk4"] Oct 13 13:45:04 crc kubenswrapper[4678]: I1013 13:45:04.606623 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0124058-8b2a-4a81-80ba-34fd9b93fd2b" path="/var/lib/kubelet/pods/e0124058-8b2a-4a81-80ba-34fd9b93fd2b/volumes" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.805559 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-brzdq/must-gather-jb7xf"] Oct 13 13:45:05 crc kubenswrapper[4678]: E1013 13:45:05.808109 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7d1823-af07-4cea-9059-23707cbffcb1" containerName="collect-profiles" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.808129 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7d1823-af07-4cea-9059-23707cbffcb1" containerName="collect-profiles" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.808718 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7d1823-af07-4cea-9059-23707cbffcb1" containerName="collect-profiles" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.811487 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.815489 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-brzdq"/"openshift-service-ca.crt" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.815708 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-brzdq"/"kube-root-ca.crt" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.815854 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-brzdq"/"default-dockercfg-xq8lf" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.838480 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-brzdq/must-gather-jb7xf"] Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.948660 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:05 crc kubenswrapper[4678]: I1013 13:45:05.948970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmkb8\" (UniqueName: \"kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.050803 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.050976 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmkb8\" (UniqueName: \"kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.051507 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.076774 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmkb8\" (UniqueName: \"kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8\") pod \"must-gather-jb7xf\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.157490 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:45:06 crc kubenswrapper[4678]: I1013 13:45:06.638738 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-brzdq/must-gather-jb7xf"] Oct 13 13:45:06 crc kubenswrapper[4678]: W1013 13:45:06.653089 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda244d6fa_08ba_4597_b7fb_75a9867d05dd.slice/crio-2b417682a9cc53b863f2fbe2c6f65f756953be98ce1cdda8421d7e9f4f39037d WatchSource:0}: Error finding container 2b417682a9cc53b863f2fbe2c6f65f756953be98ce1cdda8421d7e9f4f39037d: Status 404 returned error can't find the container with id 2b417682a9cc53b863f2fbe2c6f65f756953be98ce1cdda8421d7e9f4f39037d Oct 13 13:45:07 crc kubenswrapper[4678]: I1013 13:45:07.318379 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/must-gather-jb7xf" event={"ID":"a244d6fa-08ba-4597-b7fb-75a9867d05dd","Type":"ContainerStarted","Data":"70f96dfc2199a27b4c639582a41603ea889e836ede8b47995dc66f51b48023e1"} Oct 13 13:45:07 crc kubenswrapper[4678]: I1013 13:45:07.318735 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/must-gather-jb7xf" event={"ID":"a244d6fa-08ba-4597-b7fb-75a9867d05dd","Type":"ContainerStarted","Data":"a78cbb8920aa8a8ee54cc18dea1c1c39384841137c0afd138b6224c7a73c8933"} Oct 13 13:45:07 crc kubenswrapper[4678]: I1013 13:45:07.318753 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/must-gather-jb7xf" event={"ID":"a244d6fa-08ba-4597-b7fb-75a9867d05dd","Type":"ContainerStarted","Data":"2b417682a9cc53b863f2fbe2c6f65f756953be98ce1cdda8421d7e9f4f39037d"} Oct 13 13:45:07 crc kubenswrapper[4678]: I1013 13:45:07.339040 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-brzdq/must-gather-jb7xf" podStartSLOduration=2.3390208169999998 podStartE2EDuration="2.339020817s" podCreationTimestamp="2025-10-13 13:45:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:45:07.331022409 +0000 UTC m=+3695.415560313" watchObservedRunningTime="2025-10-13 13:45:07.339020817 +0000 UTC m=+3695.423558701" Oct 13 13:45:09 crc kubenswrapper[4678]: E1013 13:45:09.455622 4678 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.30:47146->38.102.83.30:36753: write tcp 38.102.83.30:47146->38.102.83.30:36753: write: broken pipe Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.398636 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-brzdq/crc-debug-m22vp"] Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.399808 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.537718 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.538735 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrxfp\" (UniqueName: \"kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.593455 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:45:10 crc kubenswrapper[4678]: E1013 13:45:10.593847 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.640574 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.640890 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrxfp\" (UniqueName: \"kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.640712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.666746 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrxfp\" (UniqueName: \"kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp\") pod \"crc-debug-m22vp\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: I1013 13:45:10.721604 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:10 crc kubenswrapper[4678]: W1013 13:45:10.766241 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19ff5650_db3a_49a3_a397_b8ce8078a7d1.slice/crio-6e479f0bb8a3a3e440dd64c828421ef4da90bf3fc054ec1adb9560765072b518 WatchSource:0}: Error finding container 6e479f0bb8a3a3e440dd64c828421ef4da90bf3fc054ec1adb9560765072b518: Status 404 returned error can't find the container with id 6e479f0bb8a3a3e440dd64c828421ef4da90bf3fc054ec1adb9560765072b518 Oct 13 13:45:11 crc kubenswrapper[4678]: I1013 13:45:11.352243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-m22vp" event={"ID":"19ff5650-db3a-49a3-a397-b8ce8078a7d1","Type":"ContainerStarted","Data":"7b30e121ceff4356063afa31588a29ba04f2a1ba554c8730eee302c764f8ff93"} Oct 13 13:45:11 crc kubenswrapper[4678]: I1013 13:45:11.353036 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-m22vp" event={"ID":"19ff5650-db3a-49a3-a397-b8ce8078a7d1","Type":"ContainerStarted","Data":"6e479f0bb8a3a3e440dd64c828421ef4da90bf3fc054ec1adb9560765072b518"} Oct 13 13:45:11 crc kubenswrapper[4678]: I1013 13:45:11.370906 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-brzdq/crc-debug-m22vp" podStartSLOduration=1.3708891140000001 podStartE2EDuration="1.370889114s" podCreationTimestamp="2025-10-13 13:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-13 13:45:11.367485326 +0000 UTC m=+3699.452023210" watchObservedRunningTime="2025-10-13 13:45:11.370889114 +0000 UTC m=+3699.455427008" Oct 13 13:45:22 crc kubenswrapper[4678]: I1013 13:45:22.600066 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:45:22 crc kubenswrapper[4678]: E1013 13:45:22.600944 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:45:28 crc kubenswrapper[4678]: I1013 13:45:28.797554 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:28 crc kubenswrapper[4678]: I1013 13:45:28.802687 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:28 crc kubenswrapper[4678]: I1013 13:45:28.808611 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.003427 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.003511 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.003572 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9vft\" (UniqueName: \"kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.004240 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.006237 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.015905 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.105390 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.105479 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9vft\" (UniqueName: \"kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.105564 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.106077 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.106084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.126699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9vft\" (UniqueName: \"kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft\") pod \"certified-operators-7bhr7\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.174979 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.207196 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.207265 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmccs\" (UniqueName: \"kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.207293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.310726 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.311039 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmccs\" (UniqueName: \"kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.311080 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.311178 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.311450 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.344638 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmccs\" (UniqueName: \"kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs\") pod \"community-operators-lpcsp\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.629020 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:29 crc kubenswrapper[4678]: I1013 13:45:29.881941 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:30 crc kubenswrapper[4678]: W1013 13:45:30.172846 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f860b56_a813_4f08_a696_52feb1ff297a.slice/crio-eb60400161341e750000cb946f4e4fef5647afc385ecd1b5ed6372db12b48f10 WatchSource:0}: Error finding container eb60400161341e750000cb946f4e4fef5647afc385ecd1b5ed6372db12b48f10: Status 404 returned error can't find the container with id eb60400161341e750000cb946f4e4fef5647afc385ecd1b5ed6372db12b48f10 Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.174112 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.573071 4678 generic.go:334] "Generic (PLEG): container finished" podID="1f860b56-a813-4f08-a696-52feb1ff297a" containerID="a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236" exitCode=0 Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.573407 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerDied","Data":"a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236"} Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.573443 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerStarted","Data":"eb60400161341e750000cb946f4e4fef5647afc385ecd1b5ed6372db12b48f10"} Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.575419 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.577756 4678 generic.go:334] "Generic (PLEG): container finished" podID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerID="fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62" exitCode=0 Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.577813 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerDied","Data":"fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62"} Oct 13 13:45:30 crc kubenswrapper[4678]: I1013 13:45:30.577840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerStarted","Data":"e0c4b2c209512bf35cdeed0ce5049bc2ea23b8a76fc7ad31c6e2a4a27fbe2104"} Oct 13 13:45:31 crc kubenswrapper[4678]: I1013 13:45:31.593548 4678 generic.go:334] "Generic (PLEG): container finished" podID="1f860b56-a813-4f08-a696-52feb1ff297a" containerID="2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a" exitCode=0 Oct 13 13:45:31 crc kubenswrapper[4678]: I1013 13:45:31.593675 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerDied","Data":"2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a"} Oct 13 13:45:32 crc kubenswrapper[4678]: I1013 13:45:32.608123 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerStarted","Data":"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421"} Oct 13 13:45:32 crc kubenswrapper[4678]: I1013 13:45:32.615015 4678 generic.go:334] "Generic (PLEG): container finished" podID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerID="ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630" exitCode=0 Oct 13 13:45:32 crc kubenswrapper[4678]: I1013 13:45:32.615088 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerDied","Data":"ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630"} Oct 13 13:45:32 crc kubenswrapper[4678]: I1013 13:45:32.627123 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lpcsp" podStartSLOduration=3.022865111 podStartE2EDuration="4.627103986s" podCreationTimestamp="2025-10-13 13:45:28 +0000 UTC" firstStartedPulling="2025-10-13 13:45:30.575102547 +0000 UTC m=+3718.659640431" lastFinishedPulling="2025-10-13 13:45:32.179341422 +0000 UTC m=+3720.263879306" observedRunningTime="2025-10-13 13:45:32.625109084 +0000 UTC m=+3720.709646968" watchObservedRunningTime="2025-10-13 13:45:32.627103986 +0000 UTC m=+3720.711641870" Oct 13 13:45:33 crc kubenswrapper[4678]: I1013 13:45:33.593163 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:45:33 crc kubenswrapper[4678]: E1013 13:45:33.593852 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:45:33 crc kubenswrapper[4678]: I1013 13:45:33.629706 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerStarted","Data":"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d"} Oct 13 13:45:33 crc kubenswrapper[4678]: I1013 13:45:33.649020 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7bhr7" podStartSLOduration=3.077996626 podStartE2EDuration="5.649004432s" podCreationTimestamp="2025-10-13 13:45:28 +0000 UTC" firstStartedPulling="2025-10-13 13:45:30.579324747 +0000 UTC m=+3718.663862631" lastFinishedPulling="2025-10-13 13:45:33.150332553 +0000 UTC m=+3721.234870437" observedRunningTime="2025-10-13 13:45:33.64352046 +0000 UTC m=+3721.728058344" watchObservedRunningTime="2025-10-13 13:45:33.649004432 +0000 UTC m=+3721.733542326" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.175143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.175718 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.221699 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.629505 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.629790 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.677840 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.735546 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:39 crc kubenswrapper[4678]: I1013 13:45:39.742552 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:41 crc kubenswrapper[4678]: I1013 13:45:41.052424 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:41 crc kubenswrapper[4678]: I1013 13:45:41.706312 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7bhr7" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="registry-server" containerID="cri-o://24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d" gracePeriod=2 Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.053349 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.053592 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lpcsp" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="registry-server" containerID="cri-o://5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421" gracePeriod=2 Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.530459 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.638247 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.684700 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities\") pod \"1f860b56-a813-4f08-a696-52feb1ff297a\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.684844 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content\") pod \"1f860b56-a813-4f08-a696-52feb1ff297a\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.684944 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmccs\" (UniqueName: \"kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs\") pod \"1f860b56-a813-4f08-a696-52feb1ff297a\" (UID: \"1f860b56-a813-4f08-a696-52feb1ff297a\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.687687 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities" (OuterVolumeSpecName: "utilities") pod "1f860b56-a813-4f08-a696-52feb1ff297a" (UID: "1f860b56-a813-4f08-a696-52feb1ff297a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.702568 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs" (OuterVolumeSpecName: "kube-api-access-mmccs") pod "1f860b56-a813-4f08-a696-52feb1ff297a" (UID: "1f860b56-a813-4f08-a696-52feb1ff297a"). InnerVolumeSpecName "kube-api-access-mmccs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.720590 4678 generic.go:334] "Generic (PLEG): container finished" podID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerID="24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d" exitCode=0 Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.720649 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerDied","Data":"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d"} Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.720675 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bhr7" event={"ID":"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6","Type":"ContainerDied","Data":"e0c4b2c209512bf35cdeed0ce5049bc2ea23b8a76fc7ad31c6e2a4a27fbe2104"} Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.720691 4678 scope.go:117] "RemoveContainer" containerID="24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.720823 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bhr7" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.726376 4678 generic.go:334] "Generic (PLEG): container finished" podID="1f860b56-a813-4f08-a696-52feb1ff297a" containerID="5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421" exitCode=0 Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.726449 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerDied","Data":"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421"} Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.726488 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpcsp" event={"ID":"1f860b56-a813-4f08-a696-52feb1ff297a","Type":"ContainerDied","Data":"eb60400161341e750000cb946f4e4fef5647afc385ecd1b5ed6372db12b48f10"} Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.726632 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpcsp" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.744922 4678 scope.go:117] "RemoveContainer" containerID="ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.760523 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f860b56-a813-4f08-a696-52feb1ff297a" (UID: "1f860b56-a813-4f08-a696-52feb1ff297a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.771990 4678 scope.go:117] "RemoveContainer" containerID="fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.791640 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities\") pod \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.791718 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9vft\" (UniqueName: \"kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft\") pod \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.791775 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content\") pod \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\" (UID: \"ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6\") " Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.792492 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.792513 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f860b56-a813-4f08-a696-52feb1ff297a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.792530 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmccs\" (UniqueName: \"kubernetes.io/projected/1f860b56-a813-4f08-a696-52feb1ff297a-kube-api-access-mmccs\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.793344 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities" (OuterVolumeSpecName: "utilities") pod "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" (UID: "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.796169 4678 scope.go:117] "RemoveContainer" containerID="24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.796611 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d\": container with ID starting with 24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d not found: ID does not exist" containerID="24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.796645 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d"} err="failed to get container status \"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d\": rpc error: code = NotFound desc = could not find container \"24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d\": container with ID starting with 24eba91cae550462420e8fa913051d9ff2a7ae51d0305caf70dc3f3835bafd3d not found: ID does not exist" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.796666 4678 scope.go:117] "RemoveContainer" containerID="ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.797009 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630\": container with ID starting with ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630 not found: ID does not exist" containerID="ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.797081 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630"} err="failed to get container status \"ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630\": rpc error: code = NotFound desc = could not find container \"ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630\": container with ID starting with ff6ce29f3d45b9dc1f77339390137edd8992a4ea726f5dfcb91062da1d81a630 not found: ID does not exist" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.797108 4678 scope.go:117] "RemoveContainer" containerID="fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.797352 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft" (OuterVolumeSpecName: "kube-api-access-x9vft") pod "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" (UID: "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6"). InnerVolumeSpecName "kube-api-access-x9vft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.797436 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62\": container with ID starting with fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62 not found: ID does not exist" containerID="fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.797480 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62"} err="failed to get container status \"fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62\": rpc error: code = NotFound desc = could not find container \"fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62\": container with ID starting with fc50e7ea2a6a07bff1e2bc9b17496aefae1c39f32069f9cade36622242b36f62 not found: ID does not exist" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.797508 4678 scope.go:117] "RemoveContainer" containerID="5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.825841 4678 scope.go:117] "RemoveContainer" containerID="2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.854798 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" (UID: "ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.894593 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.894638 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9vft\" (UniqueName: \"kubernetes.io/projected/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-kube-api-access-x9vft\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.894648 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.897876 4678 scope.go:117] "RemoveContainer" containerID="a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.933988 4678 scope.go:117] "RemoveContainer" containerID="5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.934594 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421\": container with ID starting with 5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421 not found: ID does not exist" containerID="5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.934646 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421"} err="failed to get container status \"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421\": rpc error: code = NotFound desc = could not find container \"5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421\": container with ID starting with 5ea5c01fcdf0fbcde2efbda622a08acec21061edc63bf76f46884a177eabe421 not found: ID does not exist" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.934680 4678 scope.go:117] "RemoveContainer" containerID="2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.935123 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a\": container with ID starting with 2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a not found: ID does not exist" containerID="2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.935221 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a"} err="failed to get container status \"2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a\": rpc error: code = NotFound desc = could not find container \"2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a\": container with ID starting with 2bcb6c669183d78a22b3477ecadb0da09e8f4b806ee7a74350c5b55aad325e8a not found: ID does not exist" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.935305 4678 scope.go:117] "RemoveContainer" containerID="a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236" Oct 13 13:45:42 crc kubenswrapper[4678]: E1013 13:45:42.938033 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236\": container with ID starting with a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236 not found: ID does not exist" containerID="a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236" Oct 13 13:45:42 crc kubenswrapper[4678]: I1013 13:45:42.938093 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236"} err="failed to get container status \"a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236\": rpc error: code = NotFound desc = could not find container \"a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236\": container with ID starting with a092550886032cb8b38dac2b5b4a3d2f9e9ed186bfacf602c72e5d297e5a2236 not found: ID does not exist" Oct 13 13:45:43 crc kubenswrapper[4678]: I1013 13:45:43.062736 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:43 crc kubenswrapper[4678]: I1013 13:45:43.073082 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7bhr7"] Oct 13 13:45:43 crc kubenswrapper[4678]: I1013 13:45:43.080993 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:43 crc kubenswrapper[4678]: I1013 13:45:43.095256 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lpcsp"] Oct 13 13:45:44 crc kubenswrapper[4678]: I1013 13:45:44.603923 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" path="/var/lib/kubelet/pods/1f860b56-a813-4f08-a696-52feb1ff297a/volumes" Oct 13 13:45:44 crc kubenswrapper[4678]: I1013 13:45:44.604565 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" path="/var/lib/kubelet/pods/ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6/volumes" Oct 13 13:45:44 crc kubenswrapper[4678]: I1013 13:45:44.746201 4678 generic.go:334] "Generic (PLEG): container finished" podID="19ff5650-db3a-49a3-a397-b8ce8078a7d1" containerID="7b30e121ceff4356063afa31588a29ba04f2a1ba554c8730eee302c764f8ff93" exitCode=0 Oct 13 13:45:44 crc kubenswrapper[4678]: I1013 13:45:44.746237 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-m22vp" event={"ID":"19ff5650-db3a-49a3-a397-b8ce8078a7d1","Type":"ContainerDied","Data":"7b30e121ceff4356063afa31588a29ba04f2a1ba554c8730eee302c764f8ff93"} Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.884914 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.912022 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-m22vp"] Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.920655 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-m22vp"] Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.989107 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host\") pod \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.989228 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrxfp\" (UniqueName: \"kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp\") pod \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\" (UID: \"19ff5650-db3a-49a3-a397-b8ce8078a7d1\") " Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.989238 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host" (OuterVolumeSpecName: "host") pod "19ff5650-db3a-49a3-a397-b8ce8078a7d1" (UID: "19ff5650-db3a-49a3-a397-b8ce8078a7d1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:45:45 crc kubenswrapper[4678]: I1013 13:45:45.989552 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/19ff5650-db3a-49a3-a397-b8ce8078a7d1-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.002207 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp" (OuterVolumeSpecName: "kube-api-access-rrxfp") pod "19ff5650-db3a-49a3-a397-b8ce8078a7d1" (UID: "19ff5650-db3a-49a3-a397-b8ce8078a7d1"). InnerVolumeSpecName "kube-api-access-rrxfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.111740 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrxfp\" (UniqueName: \"kubernetes.io/projected/19ff5650-db3a-49a3-a397-b8ce8078a7d1-kube-api-access-rrxfp\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.593165 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:45:46 crc kubenswrapper[4678]: E1013 13:45:46.593415 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.603256 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19ff5650-db3a-49a3-a397-b8ce8078a7d1" path="/var/lib/kubelet/pods/19ff5650-db3a-49a3-a397-b8ce8078a7d1/volumes" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.804116 4678 scope.go:117] "RemoveContainer" containerID="7b30e121ceff4356063afa31588a29ba04f2a1ba554c8730eee302c764f8ff93" Oct 13 13:45:46 crc kubenswrapper[4678]: I1013 13:45:46.804169 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-m22vp" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.140837 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-brzdq/crc-debug-gnnb9"] Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141379 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141398 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141414 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="extract-content" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141422 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="extract-content" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141434 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="extract-content" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141443 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="extract-content" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141476 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19ff5650-db3a-49a3-a397-b8ce8078a7d1" containerName="container-00" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141485 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="19ff5650-db3a-49a3-a397-b8ce8078a7d1" containerName="container-00" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141501 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141510 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141530 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="extract-utilities" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141538 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="extract-utilities" Oct 13 13:45:47 crc kubenswrapper[4678]: E1013 13:45:47.141556 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="extract-utilities" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141566 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="extract-utilities" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141811 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="19ff5650-db3a-49a3-a397-b8ce8078a7d1" containerName="container-00" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141830 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae644648-c4c7-4f2b-b39d-5fa8ddaf12d6" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.141848 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f860b56-a813-4f08-a696-52feb1ff297a" containerName="registry-server" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.142668 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.232372 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4tqj\" (UniqueName: \"kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.232673 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.334090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4tqj\" (UniqueName: \"kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.334175 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.334328 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.354568 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4tqj\" (UniqueName: \"kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj\") pod \"crc-debug-gnnb9\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.463272 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.815746 4678 generic.go:334] "Generic (PLEG): container finished" podID="841871a3-cc1a-45ee-8b48-5ee9826f8dd8" containerID="41fbd5b814f00f81e30916eee8b2602c7e2df39f0af2db3493ae3f8ef9315d25" exitCode=0 Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.815869 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" event={"ID":"841871a3-cc1a-45ee-8b48-5ee9826f8dd8","Type":"ContainerDied","Data":"41fbd5b814f00f81e30916eee8b2602c7e2df39f0af2db3493ae3f8ef9315d25"} Oct 13 13:45:47 crc kubenswrapper[4678]: I1013 13:45:47.816084 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" event={"ID":"841871a3-cc1a-45ee-8b48-5ee9826f8dd8","Type":"ContainerStarted","Data":"11d1f0c3384d21f3e1b7564abfe3df0044916ba46a08a9922ba838e836fe005e"} Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.274092 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-gnnb9"] Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.283995 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-gnnb9"] Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.912330 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.963363 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host\") pod \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.963509 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host" (OuterVolumeSpecName: "host") pod "841871a3-cc1a-45ee-8b48-5ee9826f8dd8" (UID: "841871a3-cc1a-45ee-8b48-5ee9826f8dd8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.963941 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4tqj\" (UniqueName: \"kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj\") pod \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\" (UID: \"841871a3-cc1a-45ee-8b48-5ee9826f8dd8\") " Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.964395 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:48 crc kubenswrapper[4678]: I1013 13:45:48.969156 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj" (OuterVolumeSpecName: "kube-api-access-l4tqj") pod "841871a3-cc1a-45ee-8b48-5ee9826f8dd8" (UID: "841871a3-cc1a-45ee-8b48-5ee9826f8dd8"). InnerVolumeSpecName "kube-api-access-l4tqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.065699 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4tqj\" (UniqueName: \"kubernetes.io/projected/841871a3-cc1a-45ee-8b48-5ee9826f8dd8-kube-api-access-l4tqj\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.435999 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-brzdq/crc-debug-wnwrq"] Oct 13 13:45:49 crc kubenswrapper[4678]: E1013 13:45:49.436777 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841871a3-cc1a-45ee-8b48-5ee9826f8dd8" containerName="container-00" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.436791 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="841871a3-cc1a-45ee-8b48-5ee9826f8dd8" containerName="container-00" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.436971 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="841871a3-cc1a-45ee-8b48-5ee9826f8dd8" containerName="container-00" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.437622 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.474418 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.474774 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkssz\" (UniqueName: \"kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.576703 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkssz\" (UniqueName: \"kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.577001 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.577141 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.595665 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkssz\" (UniqueName: \"kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz\") pod \"crc-debug-wnwrq\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.755718 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.835370 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-gnnb9" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.835378 4678 scope.go:117] "RemoveContainer" containerID="41fbd5b814f00f81e30916eee8b2602c7e2df39f0af2db3493ae3f8ef9315d25" Oct 13 13:45:49 crc kubenswrapper[4678]: I1013 13:45:49.837453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" event={"ID":"cf8c6776-0b1f-4451-9b99-916bb0e2643a","Type":"ContainerStarted","Data":"1dc491eea984dc125c29a5b1593a1f609b14017c694eda8ec7d9eedb0fde54cf"} Oct 13 13:45:50 crc kubenswrapper[4678]: I1013 13:45:50.602977 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="841871a3-cc1a-45ee-8b48-5ee9826f8dd8" path="/var/lib/kubelet/pods/841871a3-cc1a-45ee-8b48-5ee9826f8dd8/volumes" Oct 13 13:45:50 crc kubenswrapper[4678]: I1013 13:45:50.847582 4678 generic.go:334] "Generic (PLEG): container finished" podID="cf8c6776-0b1f-4451-9b99-916bb0e2643a" containerID="63a383fda4e54c12ac236f25590c09da7fbb84c5bf8f63c8fc411cd2e7adc0bf" exitCode=0 Oct 13 13:45:50 crc kubenswrapper[4678]: I1013 13:45:50.847641 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" event={"ID":"cf8c6776-0b1f-4451-9b99-916bb0e2643a","Type":"ContainerDied","Data":"63a383fda4e54c12ac236f25590c09da7fbb84c5bf8f63c8fc411cd2e7adc0bf"} Oct 13 13:45:50 crc kubenswrapper[4678]: I1013 13:45:50.936100 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-wnwrq"] Oct 13 13:45:50 crc kubenswrapper[4678]: I1013 13:45:50.950785 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-brzdq/crc-debug-wnwrq"] Oct 13 13:45:51 crc kubenswrapper[4678]: I1013 13:45:51.970720 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.032426 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host\") pod \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.032579 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host" (OuterVolumeSpecName: "host") pod "cf8c6776-0b1f-4451-9b99-916bb0e2643a" (UID: "cf8c6776-0b1f-4451-9b99-916bb0e2643a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.032649 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkssz\" (UniqueName: \"kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz\") pod \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\" (UID: \"cf8c6776-0b1f-4451-9b99-916bb0e2643a\") " Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.033579 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cf8c6776-0b1f-4451-9b99-916bb0e2643a-host\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.037439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz" (OuterVolumeSpecName: "kube-api-access-tkssz") pod "cf8c6776-0b1f-4451-9b99-916bb0e2643a" (UID: "cf8c6776-0b1f-4451-9b99-916bb0e2643a"). InnerVolumeSpecName "kube-api-access-tkssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.135661 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkssz\" (UniqueName: \"kubernetes.io/projected/cf8c6776-0b1f-4451-9b99-916bb0e2643a-kube-api-access-tkssz\") on node \"crc\" DevicePath \"\"" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.606296 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8c6776-0b1f-4451-9b99-916bb0e2643a" path="/var/lib/kubelet/pods/cf8c6776-0b1f-4451-9b99-916bb0e2643a/volumes" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.868213 4678 scope.go:117] "RemoveContainer" containerID="63a383fda4e54c12ac236f25590c09da7fbb84c5bf8f63c8fc411cd2e7adc0bf" Oct 13 13:45:52 crc kubenswrapper[4678]: I1013 13:45:52.868245 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/crc-debug-wnwrq" Oct 13 13:45:56 crc kubenswrapper[4678]: I1013 13:45:56.160137 4678 scope.go:117] "RemoveContainer" containerID="e9356078dd51f50c6c8c99a3c3b0c2a419e303f89dbffcbf3ac71ef52e1f353d" Oct 13 13:45:59 crc kubenswrapper[4678]: I1013 13:45:59.592905 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:45:59 crc kubenswrapper[4678]: E1013 13:45:59.593938 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.133670 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd59c4d76-26t4f_aef9981e-9e19-4a3a-9f29-a82f4dfdc610/barbican-api/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.366287 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57bd8d65bb-fghbr_8a9e7c5d-3070-4d64-b750-525ec923fc9c/barbican-keystone-listener/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.383049 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd59c4d76-26t4f_aef9981e-9e19-4a3a-9f29-a82f4dfdc610/barbican-api-log/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.419530 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57bd8d65bb-fghbr_8a9e7c5d-3070-4d64-b750-525ec923fc9c/barbican-keystone-listener-log/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.578492 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6b7dd7f4b7-jwrwm_b1fd2f17-a805-4985-afba-033768f4d77f/barbican-worker/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.632130 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6b7dd7f4b7-jwrwm_b1fd2f17-a805-4985-afba-033768f4d77f/barbican-worker-log/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.773174 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2s5jf_31ca8ec6-826e-4f5b-9e58-160e18674d9f/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.868942 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/ceilometer-central-agent/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.900964 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/ceilometer-notification-agent/0.log" Oct 13 13:46:08 crc kubenswrapper[4678]: I1013 13:46:08.983862 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/proxy-httpd/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.033924 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d7efc701-cf6f-4605-b72d-65e75bd5ce29/sg-core/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.125381 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8e246acd-cb1e-42d4-935b-53b777639732/cinder-api/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.242203 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8e246acd-cb1e-42d4-935b-53b777639732/cinder-api-log/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.365728 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_af635b84-5ac7-42ee-ac5c-55a2c5e20932/probe/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.384242 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_af635b84-5ac7-42ee-ac5c-55a2c5e20932/cinder-scheduler/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.524720 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dnk2z_e0fd32ad-0b5d-4287-a5db-79c362e818e6/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.621391 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2rdzz_567d50a7-a8dd-4bdb-82dc-8bee34a93d25/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.758599 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-hp6vd_0dacde92-4685-4950-8219-d634cbd0b1cc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:09 crc kubenswrapper[4678]: I1013 13:46:09.828846 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/init/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.024342 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/init/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.054270 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-hn894_48c09433-04bb-4cd0-8bfd-72606ebb285a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.097561 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-clzsn_d9f8446d-e7a4-46b7-b62e-f4761a396473/dnsmasq-dns/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.249926 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1b7bcad6-33d5-425e-aeda-70bf6518c557/glance-httpd/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.252038 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1b7bcad6-33d5-425e-aeda-70bf6518c557/glance-log/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.449768 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0b4628b6-c114-4448-b559-79861b376761/glance-httpd/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.593159 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:46:10 crc kubenswrapper[4678]: E1013 13:46:10.593495 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.699729 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_0b4628b6-c114-4448-b559-79861b376761/glance-log/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.834556 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94756fb8d-9h8cs_5bc24a68-de18-47d6-9573-67e978a05b05/horizon/0.log" Oct 13 13:46:10 crc kubenswrapper[4678]: I1013 13:46:10.978754 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-kb7x6_3608a541-f8d9-4931-a7d1-cbf90f6c505f/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.209414 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-94756fb8d-9h8cs_5bc24a68-de18-47d6-9573-67e978a05b05/horizon-log/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.252427 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-ssndt_84dfdaf8-938e-4eca-af31-79b85a483599/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.481314 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-b4bffbf65-fc8vg_f6470c70-b6b8-456a-9ca0-5bc9c9122674/keystone-api/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.527412 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29339341-xm4f5_2a462e27-6016-43fd-8989-d97b06e1ea3c/keystone-cron/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.673761 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_20a83011-dd00-465e-97cc-0bc354fec144/kube-state-metrics/0.log" Oct 13 13:46:11 crc kubenswrapper[4678]: I1013 13:46:11.760035 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-m62j6_08cf88a5-76f1-455d-a838-9d8e37902c6e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:12 crc kubenswrapper[4678]: I1013 13:46:12.150347 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dcbfc8897-x7vwx_cc052259-a50b-4d53-ba5a-bbc8458aaa97/neutron-httpd/0.log" Oct 13 13:46:12 crc kubenswrapper[4678]: I1013 13:46:12.152636 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dcbfc8897-x7vwx_cc052259-a50b-4d53-ba5a-bbc8458aaa97/neutron-api/0.log" Oct 13 13:46:12 crc kubenswrapper[4678]: I1013 13:46:12.196469 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-xmrk6_8a30e130-e610-4241-beb7-0a5d88fda83b/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:12 crc kubenswrapper[4678]: I1013 13:46:12.755037 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6c38d115-b798-4e2c-986c-c9541832f1bb/nova-api-log/0.log" Oct 13 13:46:12 crc kubenswrapper[4678]: I1013 13:46:12.840393 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0f4da22c-d9e7-4f23-af49-a6ed4551b4b8/nova-cell0-conductor-conductor/0.log" Oct 13 13:46:13 crc kubenswrapper[4678]: I1013 13:46:13.131707 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_902f91dc-30b7-4350-9d03-9517e3b911ae/nova-cell1-conductor-conductor/0.log" Oct 13 13:46:13 crc kubenswrapper[4678]: I1013 13:46:13.270004 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5b4c17c9-33f5-406f-9e2f-2fa0eef6171d/nova-cell1-novncproxy-novncproxy/0.log" Oct 13 13:46:13 crc kubenswrapper[4678]: I1013 13:46:13.278775 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6c38d115-b798-4e2c-986c-c9541832f1bb/nova-api-api/0.log" Oct 13 13:46:13 crc kubenswrapper[4678]: I1013 13:46:13.569216 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-w28hm_204374ee-a5e4-4a02-bada-f57883e5c571/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:13 crc kubenswrapper[4678]: I1013 13:46:13.731579 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0b7bb30d-3831-467e-8067-67626163b30a/nova-metadata-log/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.000885 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_ac735c75-e8f1-402a-a7d7-d213a9badd88/nova-scheduler-scheduler/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.001440 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/mysql-bootstrap/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.186349 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/galera/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.189848 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7950a23a-e8d5-4d7e-a4d5-a30f535624b3/mysql-bootstrap/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.425600 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/mysql-bootstrap/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.582808 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/mysql-bootstrap/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.629711 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_16bc2692-fbba-4e51-92b4-805141199ee2/galera/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.769018 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_0b10ccdc-9b5e-4093-8a72-5541ad94e2f3/openstackclient/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.875810 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-dv2nq_2e724ee3-a6c7-4b75-9c60-3fe2f233b96b/ovn-controller/0.log" Oct 13 13:46:14 crc kubenswrapper[4678]: I1013 13:46:14.953285 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0b7bb30d-3831-467e-8067-67626163b30a/nova-metadata-metadata/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.038226 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-67jfd_e2f127fd-3a0a-44c7-9849-de6e42149fd7/openstack-network-exporter/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.223226 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server-init/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.406700 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.437489 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovsdb-server-init/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.442422 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4z6wg_0daece72-e6c8-427f-a970-496545b49160/ovs-vswitchd/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.658772 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_030132a0-ddd0-4001-9216-9a2962cb9f1c/openstack-network-exporter/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.700499 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lvc85_416f5c49-3fd6-4789-924a-4e0ba0279fa5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.715266 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_030132a0-ddd0-4001-9216-9a2962cb9f1c/ovn-northd/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.875730 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b1da318b-1924-4ce5-9a9a-a1b483c0ffb0/openstack-network-exporter/0.log" Oct 13 13:46:15 crc kubenswrapper[4678]: I1013 13:46:15.904882 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_b1da318b-1924-4ce5-9a9a-a1b483c0ffb0/ovsdbserver-nb/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.091088 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_352ce882-938a-4fe3-b4ce-7e01a5a750e6/openstack-network-exporter/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.155757 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_352ce882-938a-4fe3-b4ce-7e01a5a750e6/ovsdbserver-sb/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.243936 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-694bd84568-p2blx_776b5cb1-0247-44a1-b63b-cffb9a30e588/placement-api/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.373821 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-694bd84568-p2blx_776b5cb1-0247-44a1-b63b-cffb9a30e588/placement-log/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.432328 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/setup-container/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.805221 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/setup-container/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.825913 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a1fd79d4-f2b6-4427-9061-f56b0f1bc0fe/rabbitmq/0.log" Oct 13 13:46:16 crc kubenswrapper[4678]: I1013 13:46:16.909159 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/setup-container/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.132725 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/setup-container/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.177104 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-g2bsv_b8dc18d2-df42-4546-981a-cfe13dcd879f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.218153 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_1c44f667-45a9-4e57-b8f9-ffd8da960a6e/rabbitmq/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.359860 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-s9vpt_26b2ddaa-5eaa-48c5-947b-507b97253a67/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.482629 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-9s9sw_d4c81445-8564-47a3-ab04-db591cbf1a86/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.647239 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pbbsm_4db4bd12-e81c-4bcc-8f80-ed20807d5856/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:17 crc kubenswrapper[4678]: I1013 13:46:17.755577 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2p9wv_78c20a4a-fd2b-4fa7-bf79-4ae7559b312e/ssh-known-hosts-edpm-deployment/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.083867 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f9fcfccd7-n448c_6d9ab7cf-ca73-4d0f-9338-68536f651533/proxy-server/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.086796 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6x2jw_07587cb9-f0d2-41d5-a596-7650a30f9cd1/swift-ring-rebalance/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.146140 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5f9fcfccd7-n448c_6d9ab7cf-ca73-4d0f-9338-68536f651533/proxy-httpd/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.278379 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-auditor/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.325672 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-reaper/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.440736 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-replicator/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.464698 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-auditor/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.491277 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/account-server/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.611460 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-replicator/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.628153 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-server/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.723797 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/container-updater/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.777858 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-auditor/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.812064 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-expirer/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.885516 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-replicator/0.log" Oct 13 13:46:18 crc kubenswrapper[4678]: I1013 13:46:18.950637 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-server/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.001826 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/object-updater/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.027120 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/rsync/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.164783 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e2337ebf-ba94-424a-9b45-fb68453bbf64/swift-recon-cron/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.225635 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-x56k2_d7940e67-463a-4cc5-812e-0bde560d6eda/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.442169 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_bac4fcc4-3dd9-44bc-b6cd-bf306cd8a71a/tempest-tests-tempest-tests-runner/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.452535 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_49ee646c-fa4a-4332-92d3-0c70d93e49da/test-operator-logs-container/0.log" Oct 13 13:46:19 crc kubenswrapper[4678]: I1013 13:46:19.674335 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-sqqln_5fc607e5-2f31-42b9-ba24-ba02b1516a75/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 13 13:46:23 crc kubenswrapper[4678]: I1013 13:46:23.594450 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:46:23 crc kubenswrapper[4678]: E1013 13:46:23.595597 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:46:28 crc kubenswrapper[4678]: I1013 13:46:28.187844 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_911b721e-cee9-45db-ae01-fb84bb431847/memcached/0.log" Oct 13 13:46:34 crc kubenswrapper[4678]: I1013 13:46:34.592612 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:46:34 crc kubenswrapper[4678]: E1013 13:46:34.593377 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:46:43 crc kubenswrapper[4678]: I1013 13:46:43.912596 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.065704 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.100506 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.132439 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.290442 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/extract/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.312781 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/util/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.366326 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_846181667fa6ccacf1d0451aa593d9923f2a5602c1dd75bdb27f69fe846hc7m_0908157e-2110-4b3d-baa9-92a9d8c7a3c4/pull/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.500822 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-srpr5_efab1e79-db75-4c28-9046-f5fb22a49c76/kube-rbac-proxy/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.538227 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-srpr5_efab1e79-db75-4c28-9046-f5fb22a49c76/manager/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.596112 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-hcblt_7809e838-a0e6-4bc7-828f-bc6aaa4ebca5/kube-rbac-proxy/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.754423 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-hcblt_7809e838-a0e6-4bc7-828f-bc6aaa4ebca5/manager/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.792427 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ltlvn_7a378ef8-b17b-465d-aabb-e289cd9719cf/manager/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.824996 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-ltlvn_7a378ef8-b17b-465d-aabb-e289cd9719cf/kube-rbac-proxy/0.log" Oct 13 13:46:44 crc kubenswrapper[4678]: I1013 13:46:44.962429 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-wf2p9_93f6cd1f-8a49-4a2b-927a-0c4a5654891d/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.047192 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-wf2p9_93f6cd1f-8a49-4a2b-927a-0c4a5654891d/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.147637 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-c74gx_2925bcf7-ade2-4da7-a99e-3d187eafd11a/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.148089 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-c74gx_2925bcf7-ade2-4da7-a99e-3d187eafd11a/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.228776 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-xg5q2_421e19af-df3c-4736-a6d1-ba43a6f8856d/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.335908 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-xg5q2_421e19af-df3c-4736-a6d1-ba43a6f8856d/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.389088 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-xb7kd_cc1d3202-a214-4cbb-a50a-7d3ecb20409f/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.568086 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vfc96_11cc73f1-36f5-4221-9b67-fb4e7611ae65/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.581153 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-xb7kd_cc1d3202-a214-4cbb-a50a-7d3ecb20409f/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.614370 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vfc96_11cc73f1-36f5-4221-9b67-fb4e7611ae65/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.757313 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-qwthm_c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.815199 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-qwthm_c5d4ef54-8aad-4cf4-9d10-97f3ceb91df1/manager/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.916152 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-64wl2_872c1819-1912-4171-8e47-35843008daaf/kube-rbac-proxy/0.log" Oct 13 13:46:45 crc kubenswrapper[4678]: I1013 13:46:45.942497 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-64wl2_872c1819-1912-4171-8e47-35843008daaf/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.035667 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-796dd_781f6537-803f-476e-a741-2030455c777f/kube-rbac-proxy/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.107536 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-796dd_781f6537-803f-476e-a741-2030455c777f/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.185945 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-f4f4z_0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a/kube-rbac-proxy/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.268634 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-f4f4z_0d64c948-9ec5-47c2-9a6d-0db7a7e76f7a/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.342092 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-hgmrn_d89268fa-c74a-49e5-9988-e6a9fb33d5c4/kube-rbac-proxy/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.418700 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-hgmrn_d89268fa-c74a-49e5-9988-e6a9fb33d5c4/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.507595 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-h9vf6_b31e9b57-89fe-42a7-ad38-3ae06224ae58/kube-rbac-proxy/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.527310 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-h9vf6_b31e9b57-89fe-42a7-ad38-3ae06224ae58/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.635855 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z_a2deeac1-517e-4e4f-974a-1f0267868448/kube-rbac-proxy/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.696869 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dsgd2z_a2deeac1-517e-4e4f-974a-1f0267868448/manager/0.log" Oct 13 13:46:46 crc kubenswrapper[4678]: I1013 13:46:46.879209 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d8c8d84c-lsmxm_3137ba4b-0e52-456e-9951-08989c113034/kube-rbac-proxy/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.006912 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-766c4fbc4c-ddcpv_e6f72e4a-d252-4053-9665-db9af69e697e/kube-rbac-proxy/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.243343 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-766c4fbc4c-ddcpv_e6f72e4a-d252-4053-9665-db9af69e697e/operator/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.310370 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4fn9q_dadbd9ec-7f53-4555-951c-2d090c017417/registry-server/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.548081 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-5rt5t_953d626f-b1c6-492b-8473-52da7c6edc3f/kube-rbac-proxy/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.609977 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-wvf4s_b2063d32-c006-427e-b9a6-84007b0c1d22/kube-rbac-proxy/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.687793 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-5rt5t_953d626f-b1c6-492b-8473-52da7c6edc3f/manager/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.745271 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-wvf4s_b2063d32-c006-427e-b9a6-84007b0c1d22/manager/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.876965 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-ngdjz_6083ce11-ee7f-4bbc-b5ff-c2816dfe6300/operator/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.978813 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-74d8c8d84c-lsmxm_3137ba4b-0e52-456e-9951-08989c113034/manager/0.log" Oct 13 13:46:47 crc kubenswrapper[4678]: I1013 13:46:47.990875 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-v5b9q_2e600274-90bb-46a9-b41a-f642ed284900/kube-rbac-proxy/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.107149 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-v5b9q_2e600274-90bb-46a9-b41a-f642ed284900/manager/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.142081 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qzgc8_b7188e6b-9e98-477e-bbe8-778da7f9bda1/kube-rbac-proxy/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.252493 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-qzgc8_b7188e6b-9e98-477e-bbe8-778da7f9bda1/manager/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.315109 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-46shf_83789dd0-5367-44d6-8cf9-0f4bf92280b1/kube-rbac-proxy/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.362147 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-46shf_83789dd0-5367-44d6-8cf9-0f4bf92280b1/manager/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.449835 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-8dkj8_ec1fafb9-7b71-45b3-ade5-6df84a9fa568/kube-rbac-proxy/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.455926 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-8dkj8_ec1fafb9-7b71-45b3-ade5-6df84a9fa568/manager/0.log" Oct 13 13:46:48 crc kubenswrapper[4678]: I1013 13:46:48.592255 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:46:48 crc kubenswrapper[4678]: E1013 13:46:48.592514 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:02 crc kubenswrapper[4678]: I1013 13:47:02.602707 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:47:02 crc kubenswrapper[4678]: E1013 13:47:02.604009 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:03 crc kubenswrapper[4678]: I1013 13:47:03.923112 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-m99jv_ad4e211d-f191-4262-9e70-d1f4fcaddeff/control-plane-machine-set-operator/0.log" Oct 13 13:47:04 crc kubenswrapper[4678]: I1013 13:47:04.074704 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dkxs9_3da9a359-7295-4aa0-95e0-070be29e5eb9/kube-rbac-proxy/0.log" Oct 13 13:47:04 crc kubenswrapper[4678]: I1013 13:47:04.127694 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-dkxs9_3da9a359-7295-4aa0-95e0-070be29e5eb9/machine-api-operator/0.log" Oct 13 13:47:13 crc kubenswrapper[4678]: I1013 13:47:13.592754 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:47:13 crc kubenswrapper[4678]: E1013 13:47:13.594013 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:15 crc kubenswrapper[4678]: I1013 13:47:15.474844 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vw4gz_e0a73006-9a7f-41d4-a99c-d749fdfad405/cert-manager-controller/0.log" Oct 13 13:47:15 crc kubenswrapper[4678]: I1013 13:47:15.625647 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-l8c98_2fc7aa8b-8b52-4579-8a8c-7b505f9d6ae1/cert-manager-cainjector/0.log" Oct 13 13:47:15 crc kubenswrapper[4678]: I1013 13:47:15.671500 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-pczp4_2d01b76f-171e-4e27-a020-3e43c2405b19/cert-manager-webhook/0.log" Oct 13 13:47:25 crc kubenswrapper[4678]: I1013 13:47:25.592937 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:47:25 crc kubenswrapper[4678]: E1013 13:47:25.593988 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:26 crc kubenswrapper[4678]: I1013 13:47:26.582901 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-c5csj_e3884e53-e229-44e5-8d8c-42fcc654f657/nmstate-console-plugin/0.log" Oct 13 13:47:26 crc kubenswrapper[4678]: I1013 13:47:26.754260 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-rtdjd_5439ee39-b0d6-43fc-92aa-f5240c460afc/nmstate-handler/0.log" Oct 13 13:47:26 crc kubenswrapper[4678]: I1013 13:47:26.802606 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-nws2x_00b6c406-a244-403d-9ca0-2406de2da1f5/nmstate-metrics/0.log" Oct 13 13:47:26 crc kubenswrapper[4678]: I1013 13:47:26.851335 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-nws2x_00b6c406-a244-403d-9ca0-2406de2da1f5/kube-rbac-proxy/0.log" Oct 13 13:47:26 crc kubenswrapper[4678]: I1013 13:47:26.996867 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hk95b_0a00ef2f-1663-499a-97a7-1b14e7d7f3c1/nmstate-operator/0.log" Oct 13 13:47:27 crc kubenswrapper[4678]: I1013 13:47:27.040008 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-tbt2v_ae486b1a-e0b6-45ff-91f2-6bb967b88436/nmstate-webhook/0.log" Oct 13 13:47:38 crc kubenswrapper[4678]: I1013 13:47:38.592096 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:47:38 crc kubenswrapper[4678]: E1013 13:47:38.592841 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:39 crc kubenswrapper[4678]: I1013 13:47:39.830159 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-x9tg5_ae7e1bbe-3d07-40db-991b-797c25efd443/kube-rbac-proxy/0.log" Oct 13 13:47:39 crc kubenswrapper[4678]: I1013 13:47:39.925129 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-x9tg5_ae7e1bbe-3d07-40db-991b-797c25efd443/controller/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.041457 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.204466 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.210708 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.236189 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.236294 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.373708 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.428207 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.434379 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.454126 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.568415 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-frr-files/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.596837 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-reloader/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.625881 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/controller/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.636615 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/cp-metrics/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.817107 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/kube-rbac-proxy/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.824210 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/frr-metrics/0.log" Oct 13 13:47:40 crc kubenswrapper[4678]: I1013 13:47:40.847045 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/kube-rbac-proxy-frr/0.log" Oct 13 13:47:41 crc kubenswrapper[4678]: I1013 13:47:41.060432 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-xp775_61e5f2b7-9ff7-4c33-a4ff-bbc07fbea5ac/frr-k8s-webhook-server/0.log" Oct 13 13:47:41 crc kubenswrapper[4678]: I1013 13:47:41.080192 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/reloader/0.log" Oct 13 13:47:41 crc kubenswrapper[4678]: I1013 13:47:41.354244 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-797dff698d-tbgv4_51bd00ff-af4f-4405-9f25-20b4d22e7020/manager/0.log" Oct 13 13:47:41 crc kubenswrapper[4678]: I1013 13:47:41.510395 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6545fd778d-8xqwq_0b593c83-b8a6-4cf2-8dbe-245b3d3abfc5/webhook-server/0.log" Oct 13 13:47:41 crc kubenswrapper[4678]: I1013 13:47:41.589827 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mbsh9_f0f04693-55fb-40f0-8f97-c91d12a02509/kube-rbac-proxy/0.log" Oct 13 13:47:42 crc kubenswrapper[4678]: I1013 13:47:42.106854 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mbsh9_f0f04693-55fb-40f0-8f97-c91d12a02509/speaker/0.log" Oct 13 13:47:42 crc kubenswrapper[4678]: I1013 13:47:42.142764 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hdzf4_ead738f8-1d8c-491d-9c14-8cf4b535badf/frr/0.log" Oct 13 13:47:52 crc kubenswrapper[4678]: I1013 13:47:52.598638 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:47:52 crc kubenswrapper[4678]: E1013 13:47:52.600742 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.584077 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.752658 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.757826 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.759995 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.909996 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/util/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.926329 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/extract/0.log" Oct 13 13:47:53 crc kubenswrapper[4678]: I1013 13:47:53.927471 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2b8jl4_d05a6818-fca2-4ab9-8d9a-4aceb001c74c/pull/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.078697 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.222414 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.248009 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.262760 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.424475 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-content/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.429514 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/extract-utilities/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.659096 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.779866 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.783175 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-b2nmp_ab56d409-9f69-413f-b208-d236ff08ed98/registry-server/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.851162 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:47:54 crc kubenswrapper[4678]: I1013 13:47:54.889444 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.013900 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-utilities/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.118835 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/extract-content/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.231892 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.378355 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-z84r9_cb08d1b2-77d0-45bf-9a67-3b32448063a1/registry-server/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.458363 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.470457 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.484223 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.618795 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/util/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.646288 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/pull/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.657391 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chb7jh_9fa7ade0-a26e-45ea-ab14-017c299f5582/extract/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.832494 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9j9mq_2c14e376-1a85-43dc-a3d1-66e37a54fd22/marketplace-operator/0.log" Oct 13 13:47:55 crc kubenswrapper[4678]: I1013 13:47:55.837023 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.034574 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.035509 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.038648 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.222552 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-content/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.268046 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.301470 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-brf7p_7495eaa2-5ad8-4110-96a8-11d4eab8682f/registry-server/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.444481 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.632261 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.635668 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.652677 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.806549 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-utilities/0.log" Oct 13 13:47:56 crc kubenswrapper[4678]: I1013 13:47:56.832720 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/extract-content/0.log" Oct 13 13:47:57 crc kubenswrapper[4678]: I1013 13:47:57.406167 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-56qm8_2ff166cb-c0b3-4451-9666-9a27bf005741/registry-server/0.log" Oct 13 13:48:06 crc kubenswrapper[4678]: I1013 13:48:06.593495 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:48:06 crc kubenswrapper[4678]: E1013 13:48:06.594361 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:48:21 crc kubenswrapper[4678]: I1013 13:48:21.593251 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:48:21 crc kubenswrapper[4678]: E1013 13:48:21.594089 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bxtwv_openshift-machine-config-operator(8c6e2144-3790-4a4a-b47c-5b3459dacdf9)\"" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" Oct 13 13:48:33 crc kubenswrapper[4678]: I1013 13:48:33.592908 4678 scope.go:117] "RemoveContainer" containerID="474b0c7d4ffaf322c545c5b9ae91d1e5d8d98511a9023b78623cbe1c3e94a7b3" Oct 13 13:48:34 crc kubenswrapper[4678]: I1013 13:48:34.306419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" event={"ID":"8c6e2144-3790-4a4a-b47c-5b3459dacdf9","Type":"ContainerStarted","Data":"bdf10e334d5f4cc7e9b909f970029cb378ff4961893bd39d284b6d345a977110"} Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.015233 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:42 crc kubenswrapper[4678]: E1013 13:48:42.016350 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8c6776-0b1f-4451-9b99-916bb0e2643a" containerName="container-00" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.016372 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8c6776-0b1f-4451-9b99-916bb0e2643a" containerName="container-00" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.016631 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8c6776-0b1f-4451-9b99-916bb0e2643a" containerName="container-00" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.018341 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.027657 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.081914 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.082104 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.082137 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbbc\" (UniqueName: \"kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.183665 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.183725 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbbc\" (UniqueName: \"kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.183824 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.184380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.184530 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.204850 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbbc\" (UniqueName: \"kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc\") pod \"redhat-operators-cf27h\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.351240 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:42 crc kubenswrapper[4678]: I1013 13:48:42.861038 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:43 crc kubenswrapper[4678]: I1013 13:48:43.412965 4678 generic.go:334] "Generic (PLEG): container finished" podID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerID="37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb" exitCode=0 Oct 13 13:48:43 crc kubenswrapper[4678]: I1013 13:48:43.413602 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerDied","Data":"37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb"} Oct 13 13:48:43 crc kubenswrapper[4678]: I1013 13:48:43.413659 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerStarted","Data":"ddea778ea7f752cd1240a4bc6188ac71e43070939cb34a8a578aae87cd649322"} Oct 13 13:48:45 crc kubenswrapper[4678]: I1013 13:48:45.434922 4678 generic.go:334] "Generic (PLEG): container finished" podID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerID="cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5" exitCode=0 Oct 13 13:48:45 crc kubenswrapper[4678]: I1013 13:48:45.435076 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerDied","Data":"cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5"} Oct 13 13:48:46 crc kubenswrapper[4678]: I1013 13:48:46.453025 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerStarted","Data":"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2"} Oct 13 13:48:46 crc kubenswrapper[4678]: I1013 13:48:46.487961 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cf27h" podStartSLOduration=2.82508698 podStartE2EDuration="5.487926346s" podCreationTimestamp="2025-10-13 13:48:41 +0000 UTC" firstStartedPulling="2025-10-13 13:48:43.415031277 +0000 UTC m=+3911.499569161" lastFinishedPulling="2025-10-13 13:48:46.077870633 +0000 UTC m=+3914.162408527" observedRunningTime="2025-10-13 13:48:46.477116044 +0000 UTC m=+3914.561653958" watchObservedRunningTime="2025-10-13 13:48:46.487926346 +0000 UTC m=+3914.572464270" Oct 13 13:48:52 crc kubenswrapper[4678]: I1013 13:48:52.352214 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:52 crc kubenswrapper[4678]: I1013 13:48:52.352852 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:52 crc kubenswrapper[4678]: I1013 13:48:52.394401 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:52 crc kubenswrapper[4678]: I1013 13:48:52.570523 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:52 crc kubenswrapper[4678]: I1013 13:48:52.632895 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:54 crc kubenswrapper[4678]: I1013 13:48:54.548432 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cf27h" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="registry-server" containerID="cri-o://09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2" gracePeriod=2 Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.003386 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.094247 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content\") pod \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.094415 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbbc\" (UniqueName: \"kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc\") pod \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.094532 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities\") pod \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\" (UID: \"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f\") " Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.095940 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities" (OuterVolumeSpecName: "utilities") pod "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" (UID: "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.104948 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc" (OuterVolumeSpecName: "kube-api-access-kwbbc") pod "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" (UID: "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f"). InnerVolumeSpecName "kube-api-access-kwbbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.196669 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbbc\" (UniqueName: \"kubernetes.io/projected/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-kube-api-access-kwbbc\") on node \"crc\" DevicePath \"\"" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.196717 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.201507 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" (UID: "9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.346033 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.562768 4678 generic.go:334] "Generic (PLEG): container finished" podID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerID="09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2" exitCode=0 Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.562835 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerDied","Data":"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2"} Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.562882 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cf27h" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.562905 4678 scope.go:117] "RemoveContainer" containerID="09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.562887 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cf27h" event={"ID":"9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f","Type":"ContainerDied","Data":"ddea778ea7f752cd1240a4bc6188ac71e43070939cb34a8a578aae87cd649322"} Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.600741 4678 scope.go:117] "RemoveContainer" containerID="cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.604824 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.613486 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cf27h"] Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.633044 4678 scope.go:117] "RemoveContainer" containerID="37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.679622 4678 scope.go:117] "RemoveContainer" containerID="09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2" Oct 13 13:48:55 crc kubenswrapper[4678]: E1013 13:48:55.679957 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2\": container with ID starting with 09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2 not found: ID does not exist" containerID="09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.679988 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2"} err="failed to get container status \"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2\": rpc error: code = NotFound desc = could not find container \"09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2\": container with ID starting with 09e472d0e71f7348f0fa9d3454efb8b14db03843bf6fc802cb821ffdfeae40b2 not found: ID does not exist" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.680006 4678 scope.go:117] "RemoveContainer" containerID="cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5" Oct 13 13:48:55 crc kubenswrapper[4678]: E1013 13:48:55.680252 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5\": container with ID starting with cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5 not found: ID does not exist" containerID="cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.680270 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5"} err="failed to get container status \"cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5\": rpc error: code = NotFound desc = could not find container \"cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5\": container with ID starting with cfc58990b402c7297a36213ec81fd238ba07b65fcbf12e676ba84c94817ce0e5 not found: ID does not exist" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.680284 4678 scope.go:117] "RemoveContainer" containerID="37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb" Oct 13 13:48:55 crc kubenswrapper[4678]: E1013 13:48:55.680472 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb\": container with ID starting with 37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb not found: ID does not exist" containerID="37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb" Oct 13 13:48:55 crc kubenswrapper[4678]: I1013 13:48:55.680488 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb"} err="failed to get container status \"37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb\": rpc error: code = NotFound desc = could not find container \"37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb\": container with ID starting with 37ec9f2e5b99726e34a2c7372043919ba8ffdcd4d009c14d975e3eaa2892b0cb not found: ID does not exist" Oct 13 13:48:56 crc kubenswrapper[4678]: I1013 13:48:56.603906 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" path="/var/lib/kubelet/pods/9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f/volumes" Oct 13 13:49:29 crc kubenswrapper[4678]: I1013 13:49:29.910040 4678 generic.go:334] "Generic (PLEG): container finished" podID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerID="a78cbb8920aa8a8ee54cc18dea1c1c39384841137c0afd138b6224c7a73c8933" exitCode=0 Oct 13 13:49:29 crc kubenswrapper[4678]: I1013 13:49:29.910171 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-brzdq/must-gather-jb7xf" event={"ID":"a244d6fa-08ba-4597-b7fb-75a9867d05dd","Type":"ContainerDied","Data":"a78cbb8920aa8a8ee54cc18dea1c1c39384841137c0afd138b6224c7a73c8933"} Oct 13 13:49:29 crc kubenswrapper[4678]: I1013 13:49:29.911899 4678 scope.go:117] "RemoveContainer" containerID="a78cbb8920aa8a8ee54cc18dea1c1c39384841137c0afd138b6224c7a73c8933" Oct 13 13:49:30 crc kubenswrapper[4678]: I1013 13:49:30.714973 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-brzdq_must-gather-jb7xf_a244d6fa-08ba-4597-b7fb-75a9867d05dd/gather/0.log" Oct 13 13:49:40 crc kubenswrapper[4678]: I1013 13:49:40.772893 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-brzdq/must-gather-jb7xf"] Oct 13 13:49:40 crc kubenswrapper[4678]: I1013 13:49:40.773709 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-brzdq/must-gather-jb7xf" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="copy" containerID="cri-o://70f96dfc2199a27b4c639582a41603ea889e836ede8b47995dc66f51b48023e1" gracePeriod=2 Oct 13 13:49:40 crc kubenswrapper[4678]: I1013 13:49:40.792838 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-brzdq/must-gather-jb7xf"] Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.039648 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-brzdq_must-gather-jb7xf_a244d6fa-08ba-4597-b7fb-75a9867d05dd/copy/0.log" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.040506 4678 generic.go:334] "Generic (PLEG): container finished" podID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerID="70f96dfc2199a27b4c639582a41603ea889e836ede8b47995dc66f51b48023e1" exitCode=143 Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.203174 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-brzdq_must-gather-jb7xf_a244d6fa-08ba-4597-b7fb-75a9867d05dd/copy/0.log" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.203639 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.225898 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output\") pod \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.226401 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmkb8\" (UniqueName: \"kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8\") pod \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\" (UID: \"a244d6fa-08ba-4597-b7fb-75a9867d05dd\") " Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.236291 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8" (OuterVolumeSpecName: "kube-api-access-cmkb8") pod "a244d6fa-08ba-4597-b7fb-75a9867d05dd" (UID: "a244d6fa-08ba-4597-b7fb-75a9867d05dd"). InnerVolumeSpecName "kube-api-access-cmkb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.328912 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmkb8\" (UniqueName: \"kubernetes.io/projected/a244d6fa-08ba-4597-b7fb-75a9867d05dd-kube-api-access-cmkb8\") on node \"crc\" DevicePath \"\"" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.388808 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a244d6fa-08ba-4597-b7fb-75a9867d05dd" (UID: "a244d6fa-08ba-4597-b7fb-75a9867d05dd"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:49:41 crc kubenswrapper[4678]: I1013 13:49:41.430515 4678 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a244d6fa-08ba-4597-b7fb-75a9867d05dd-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 13 13:49:42 crc kubenswrapper[4678]: I1013 13:49:42.051114 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-brzdq_must-gather-jb7xf_a244d6fa-08ba-4597-b7fb-75a9867d05dd/copy/0.log" Oct 13 13:49:42 crc kubenswrapper[4678]: I1013 13:49:42.051795 4678 scope.go:117] "RemoveContainer" containerID="70f96dfc2199a27b4c639582a41603ea889e836ede8b47995dc66f51b48023e1" Oct 13 13:49:42 crc kubenswrapper[4678]: I1013 13:49:42.051868 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-brzdq/must-gather-jb7xf" Oct 13 13:49:42 crc kubenswrapper[4678]: I1013 13:49:42.074648 4678 scope.go:117] "RemoveContainer" containerID="a78cbb8920aa8a8ee54cc18dea1c1c39384841137c0afd138b6224c7a73c8933" Oct 13 13:49:42 crc kubenswrapper[4678]: I1013 13:49:42.616791 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" path="/var/lib/kubelet/pods/a244d6fa-08ba-4597-b7fb-75a9867d05dd/volumes" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.879176 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:26 crc kubenswrapper[4678]: E1013 13:50:26.880475 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="registry-server" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880493 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="registry-server" Oct 13 13:50:26 crc kubenswrapper[4678]: E1013 13:50:26.880517 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="extract-content" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880527 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="extract-content" Oct 13 13:50:26 crc kubenswrapper[4678]: E1013 13:50:26.880560 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="copy" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880568 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="copy" Oct 13 13:50:26 crc kubenswrapper[4678]: E1013 13:50:26.880587 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="gather" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880595 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="gather" Oct 13 13:50:26 crc kubenswrapper[4678]: E1013 13:50:26.880615 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="extract-utilities" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880624 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="extract-utilities" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880828 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="copy" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880847 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a244d6fa-08ba-4597-b7fb-75a9867d05dd" containerName="gather" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.880871 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d9d88d3-5fbc-4df1-9d0c-b33c491a3d8f" containerName="registry-server" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.882146 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.895107 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.965261 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.965316 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:26 crc kubenswrapper[4678]: I1013 13:50:26.965719 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5c89\" (UniqueName: \"kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.067308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.067374 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.067508 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5c89\" (UniqueName: \"kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.067828 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.067864 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.085828 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5c89\" (UniqueName: \"kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89\") pod \"redhat-marketplace-kqmx7\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:27 crc kubenswrapper[4678]: I1013 13:50:27.210781 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:28 crc kubenswrapper[4678]: I1013 13:50:28.224509 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:28 crc kubenswrapper[4678]: W1013 13:50:28.306479 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bcd777f_e0cc_4eb1_b117_49ed05495558.slice/crio-80aa69b43619999d1aa7c54029cf44c3c68dd0e31ec37aa26f45613c70a8f215 WatchSource:0}: Error finding container 80aa69b43619999d1aa7c54029cf44c3c68dd0e31ec37aa26f45613c70a8f215: Status 404 returned error can't find the container with id 80aa69b43619999d1aa7c54029cf44c3c68dd0e31ec37aa26f45613c70a8f215 Oct 13 13:50:28 crc kubenswrapper[4678]: I1013 13:50:28.506235 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerStarted","Data":"80aa69b43619999d1aa7c54029cf44c3c68dd0e31ec37aa26f45613c70a8f215"} Oct 13 13:50:29 crc kubenswrapper[4678]: I1013 13:50:29.517394 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bcd777f-e0cc-4eb1-b117-49ed05495558" containerID="9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee" exitCode=0 Oct 13 13:50:29 crc kubenswrapper[4678]: I1013 13:50:29.517460 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerDied","Data":"9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee"} Oct 13 13:50:30 crc kubenswrapper[4678]: I1013 13:50:30.527794 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerStarted","Data":"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd"} Oct 13 13:50:31 crc kubenswrapper[4678]: I1013 13:50:31.541154 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bcd777f-e0cc-4eb1-b117-49ed05495558" containerID="d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd" exitCode=0 Oct 13 13:50:31 crc kubenswrapper[4678]: I1013 13:50:31.541203 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerDied","Data":"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd"} Oct 13 13:50:31 crc kubenswrapper[4678]: I1013 13:50:31.544614 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 13 13:50:32 crc kubenswrapper[4678]: I1013 13:50:32.551304 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerStarted","Data":"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a"} Oct 13 13:50:32 crc kubenswrapper[4678]: I1013 13:50:32.571716 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kqmx7" podStartSLOduration=4.118230501 podStartE2EDuration="6.571700557s" podCreationTimestamp="2025-10-13 13:50:26 +0000 UTC" firstStartedPulling="2025-10-13 13:50:29.52113477 +0000 UTC m=+4017.605672654" lastFinishedPulling="2025-10-13 13:50:31.974604836 +0000 UTC m=+4020.059142710" observedRunningTime="2025-10-13 13:50:32.571222434 +0000 UTC m=+4020.655760328" watchObservedRunningTime="2025-10-13 13:50:32.571700557 +0000 UTC m=+4020.656238441" Oct 13 13:50:37 crc kubenswrapper[4678]: I1013 13:50:37.211168 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:37 crc kubenswrapper[4678]: I1013 13:50:37.211763 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:37 crc kubenswrapper[4678]: I1013 13:50:37.261279 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:37 crc kubenswrapper[4678]: I1013 13:50:37.636479 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:37 crc kubenswrapper[4678]: I1013 13:50:37.683898 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:39 crc kubenswrapper[4678]: I1013 13:50:39.612827 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kqmx7" podUID="4bcd777f-e0cc-4eb1-b117-49ed05495558" containerName="registry-server" containerID="cri-o://69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a" gracePeriod=2 Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.100425 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.157600 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities\") pod \"4bcd777f-e0cc-4eb1-b117-49ed05495558\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.157782 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content\") pod \"4bcd777f-e0cc-4eb1-b117-49ed05495558\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.158075 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5c89\" (UniqueName: \"kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89\") pod \"4bcd777f-e0cc-4eb1-b117-49ed05495558\" (UID: \"4bcd777f-e0cc-4eb1-b117-49ed05495558\") " Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.158805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities" (OuterVolumeSpecName: "utilities") pod "4bcd777f-e0cc-4eb1-b117-49ed05495558" (UID: "4bcd777f-e0cc-4eb1-b117-49ed05495558"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.165106 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89" (OuterVolumeSpecName: "kube-api-access-m5c89") pod "4bcd777f-e0cc-4eb1-b117-49ed05495558" (UID: "4bcd777f-e0cc-4eb1-b117-49ed05495558"). InnerVolumeSpecName "kube-api-access-m5c89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.175292 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bcd777f-e0cc-4eb1-b117-49ed05495558" (UID: "4bcd777f-e0cc-4eb1-b117-49ed05495558"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.260530 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.260572 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5c89\" (UniqueName: \"kubernetes.io/projected/4bcd777f-e0cc-4eb1-b117-49ed05495558-kube-api-access-m5c89\") on node \"crc\" DevicePath \"\"" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.260583 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bcd777f-e0cc-4eb1-b117-49ed05495558-utilities\") on node \"crc\" DevicePath \"\"" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.646276 4678 generic.go:334] "Generic (PLEG): container finished" podID="4bcd777f-e0cc-4eb1-b117-49ed05495558" containerID="69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a" exitCode=0 Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.646399 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerDied","Data":"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a"} Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.646447 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kqmx7" event={"ID":"4bcd777f-e0cc-4eb1-b117-49ed05495558","Type":"ContainerDied","Data":"80aa69b43619999d1aa7c54029cf44c3c68dd0e31ec37aa26f45613c70a8f215"} Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.646477 4678 scope.go:117] "RemoveContainer" containerID="69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.646734 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kqmx7" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.675306 4678 scope.go:117] "RemoveContainer" containerID="d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.676124 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.685347 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kqmx7"] Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.806206 4678 scope.go:117] "RemoveContainer" containerID="9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.847321 4678 scope.go:117] "RemoveContainer" containerID="69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a" Oct 13 13:50:40 crc kubenswrapper[4678]: E1013 13:50:40.847964 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a\": container with ID starting with 69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a not found: ID does not exist" containerID="69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.847993 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a"} err="failed to get container status \"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a\": rpc error: code = NotFound desc = could not find container \"69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a\": container with ID starting with 69b2d43418426ef9defc6da3c1ea85d9b04ec48cb3fbcbbdb918bca7a107175a not found: ID does not exist" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.848013 4678 scope.go:117] "RemoveContainer" containerID="d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd" Oct 13 13:50:40 crc kubenswrapper[4678]: E1013 13:50:40.848472 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd\": container with ID starting with d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd not found: ID does not exist" containerID="d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.848523 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd"} err="failed to get container status \"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd\": rpc error: code = NotFound desc = could not find container \"d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd\": container with ID starting with d2e6dbeb291a6ff6949a8b1d5f05babba580ed4ba2223f4d54ccfbf5e6e4b4fd not found: ID does not exist" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.848554 4678 scope.go:117] "RemoveContainer" containerID="9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee" Oct 13 13:50:40 crc kubenswrapper[4678]: E1013 13:50:40.849025 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee\": container with ID starting with 9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee not found: ID does not exist" containerID="9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee" Oct 13 13:50:40 crc kubenswrapper[4678]: I1013 13:50:40.849071 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee"} err="failed to get container status \"9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee\": rpc error: code = NotFound desc = could not find container \"9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee\": container with ID starting with 9f0ec87dea48b8cfc26daa3af902818e8a1913075219b2d2d1b5a9a74baddfee not found: ID does not exist" Oct 13 13:50:42 crc kubenswrapper[4678]: I1013 13:50:42.624838 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bcd777f-e0cc-4eb1-b117-49ed05495558" path="/var/lib/kubelet/pods/4bcd777f-e0cc-4eb1-b117-49ed05495558/volumes" Oct 13 13:50:55 crc kubenswrapper[4678]: I1013 13:50:55.506373 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bxtwv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 13 13:50:55 crc kubenswrapper[4678]: I1013 13:50:55.507138 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bxtwv" podUID="8c6e2144-3790-4a4a-b47c-5b3459dacdf9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073201727024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073201730017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073171434016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073171434015462 5ustar corecore